MACHINE VISION-BASED METHOD AND SYSTEM FOR DETERMINING A RANGE OF MOTION OF A JOINT OF A HAND OF A SUBJECT

Information

  • Patent Application
  • 20250213141
  • Publication Number
    20250213141
  • Date Filed
    December 28, 2023
    a year ago
  • Date Published
    July 03, 2025
    22 days ago
Abstract
The present disclosure provides a machine vision-based method for determining a range of motion of a joint of a hand of a subject and a system for implementing the same. The provided method comprises: facilitating the subject to perform a hand movement at a preset position; capturing at least two prior-movement images of the hand when the hand is in a neutral posture before performing the hand movement; capturing at least two post-movement images of the hand when the hand is in an assessment posture after performing the hand movement; processing the captured prior-movement images to obtain a plurality of prior-movement key point positions; processing the captured post-movement images to obtain a plurality of post-movement key point positions; and calculating the range of motion of the joint based on the plurality of prior-movement key point positions and the plurality of post-movement key point positions.
Description
FIELD OF THE INVENTION

The present invention generally relates to hand function assessment. More specifically, the present invention relates to a machine vision-based method for determining the range of motion of the joints of a hand of a subject and a system for implementing the same.


BACKGROUND OF THE INVENTION

The hand is one of the most complex and versatile anatomical structures of the human appendages. It serves a crucial function for interacting with the external environment and plays a significant role in human life. Functional disorders of the hand can limit individual independence and affect the quality of life. The determination of range of motion (ROM) is an essential component of hand function assessment. In clinical settings, hand joint angle measurement methods are commonly used to obtain the ROM of hand joints for the following reasons. Firstly, hand joint angles reflect hand function and motor ability. By measuring hand joint angles, one can understand the ROM, flexibility, and coordination of the hand joints, aiding in the assessment of hand function recovery and the extent of damage. Secondly, measuring hand joint angles can provide important evidence for clinical diagnosis and assist in rehabilitation therapy. Joint angle measurements help physicians quickly locate the site of lesions, determine the degree of damage, and formulate appropriate treatment plans. Lastly, in the field of medicine and sports science, hand joint measurements can reveal the dynamic characteristics and movement patterns of hand motion. This is crucial for understanding the biomechanical features of hand motion, optimizing motor skills, and improving sports training techniques.


Medical professionals often use universal goniometers or inclinometers to manually measure the declination angles of finger joints to assess the joint movement range. However, one significant challenge in manual assessment is intra- and inter-rater reliability. The methods involving manual assessment are also inefficient and expensive. In other approaches, electronic wearable devices were introduced to sense the hand movement. However, one of the challenges in these approaches is that the wearable devices require physical contact with the finger to achieve the best accuracy. Injuries, such as burns, wounds, lacerations or even dermatological conditions, can cause difficulties in conducting the measurement.


Adapting optical measurement systems or computer vision-based approaches provides a non-contact form of measurement. For example, Chinese patent application publication no. CN106355598A disclosed an automatic wrist and finger joint motion degree measurement method based on Kinect depth images. It is still desirable to have a more versatile vision-based hand joint angle measurement system which can measure hand joint angle for various hand movements and provide a more comprehensive hand function assessment.


SUMMARY OF THE INVENTION

According to one aspect of the present invention, a machine vision-based method for determining the range of motion of the joints of a hand of a subject is provided. The provided method comprises: facilitating the subject to perform a hand movement at a preset position; capturing at least two prior-movement images of the hand when the hand is in a neutral posture before performing the hand movement; capturing at least two post-movement images of the hand when the hand is in an assessment posture after performing the hand movement; processing the captured prior-movement images to obtain a plurality of prior-movement key point positions; processing the captured post-movement images to obtain a plurality of post-movement key point positions; and calculating the range of motion of the joint based on the plurality of prior-movement key point positions and the plurality of post-movement key point positions.


According to another aspect of the present invention, a machine vision-based system for determining a ROM of a joint of a hand of a subject is provided. The provided system comprises: a supporter configured to facilitate the subject to perform a hand movement at a preset position; at least two cameras configured to: capture at least two prior-movement images of the hand when the hand is in a neutral posture before performing the hand movement; and capture at least two post-movement images of the hand when the hand is in an assessment posture after performing the hand movement; and a processor configured to: process the captured prior-movement images to obtain a plurality of prior-movement key point positions; process the captured post-movement images to obtain a plurality of post-movement key point positions; an calculate the range of motion of the joint based on the plurality of prior-movement key point positions and the plurality of post-movement key point positions.


According to a further aspect of the present invention, a non-transitory computer-readable storage medium is provided to store a program including instructions for performing a machine vision-based method for determining a range of motion of a joint of a hand of a subject. The method comprises: facilitating the subject to perform a hand movement at a preset position; capturing at least two prior-movement images of the hand when the hand is in a neutral posture before performing the hand movement; capturing at least two post-movement images of the hand when the hand is in an assessment posture after performing the hand movement; processing the captured prior-movement images to obtain a plurality of prior-movement key point positions; processing the captured post-movement images to obtain a plurality of post-movement key point positions; and calculating the range of motion of the joint based on the plurality of prior-movement key point positions and the plurality of post-movement key point positions.


The present invention introduces an unconstrained and non-invasive machine vision-based approach, eliminating the need for direct hand contact. Unlike conventional manual measurement techniques, the present invention prevents human biases, reducing errors for more dependable results. The provided measurement method expedites hand joint angle measurements while concurrently assessing multiple hand joints. This efficiency not only saves time and resources but also facilitates comprehensive hand mobility assessments. It preserves the natural fluidity of movement by enabling unrestricted hand motion, setting it apart from wearables, gloves, or external devices. The machine vision technology closely mimics real hand activity, enhancing measurement authenticity and accuracy.


The invention also offers timely rehabilitation progress feedback, aiding therapy efforts. Furthermore, its automated data processing generates detailed angle ranges, statistical insights, and visual representations, simplifying data interpretation and analysis for healthcare professionals. It also offers automated tracking of finger and upper limb movements, delivering precise measurements and insights into an individual's fine motor skills. This capability proves invaluable for therapists assessing patient recovery and evaluating fine motor skills in professions that demand such precision. Beyond its immediate applications, this advancement holds significant promise for research in biomechanics, neuroscience, and human-computer interaction.





BRIEF DESCRIPTION OF THE DRAWINGS

Aspects of the present disclosure may be readily understood from the following detailed description with reference to the accompanying figures. The illustrations may not necessarily be drawn to scale. That is, the dimensions of the various features may be arbitrarily increased or reduced for clarity of discussion. There may be distinctions between the artistic renditions in the present disclosure and the actual apparatus due to manufacturing processes and tolerances. Common reference numerals may be used throughout the drawings and the detailed description to indicate the same or similar components.



FIG. 1 shows a simplified diagram of a machine vision-based system for determining a range of motions of joints of a hand of a subject in accordance with one embodiment of the present invention.



FIG. 2 shows a flow chart of a machine vision-based method for determining a range of motions of joints of a hand of a subject in accordance with one embodiment of the present invention.



FIG. 3 shows more details on how prior-movement/post movement key point positions are obtained for calculating range of motion.



FIG. 4. shows how mapping of 2D positions from the prior-movement/post-movement images into the 3D space can be regarded as a triangulation problem.



FIG. 5A shows an anatomical model of the hand and FIG. 5B shows a kinematic model used in determining the range of motion.



FIGS. 6A to 6G show seven hand movements respectively designed for hand function assessment.



FIG. 7 shows how a plurality of 3D post-movement key point positions is projected on a plane perpendicular to the palm plane or a plane in parallel with the palm plane to obtain the plurality of post-movement key point positions for calculating the ROM.





DETAILED DESCRIPTION

In the following description, preferred examples of the present invention will be set forth as embodiments which are to be regarded as illustrative rather than restrictive. Specific details may be omitted so as not to obscure the present disclosure; however, the disclosure is written to enable one skilled in the art to practice the teachings herein without undue experimentation.


Referring to FIG. 1 for the following description. In accordance with one embodiment of the present invention, a machine vision-based system 100 for determining a range of motions of joints of a hand of a subject. The system 100 comprises a hand supporter 101, two or more cameras 102 and a processor.


The hand supporter 101 is configured to support hands of the subject to facilitate the subject to perform a hand movement at a preset position.


The two or more cameras 102 are configured to capture images of the hand from two or more perspectives respectively. More specifically, the cameras 102 may be configured to: capture prior-movement images of the hand when the hand is in a neutral posture before performing the hand movement; and capture post-movement images of the hand when the hand has an assessment posture after performing the hand movement.


In some embodiments, each camera may be calibrated with respect to a three-dimensional space to obtain a respective camera projection matrix for reconstructing the 3D coordinates of joints from the 2D pixel coordinates in the image.


The camera projection matrix is a representation of the projecting points in the 3D space onto the pixel coordinate system. The process of projecting points in the 3D space onto the pixel coordinate system essentially involves taking global coordinates of a 3D point P (Xw,Yw,Zw), mapping them using the pinhole imaging principle to the 2D image coordinate system, and applying rigid body transformation to convert them into 2D pixel coordinates p(u,v) in the image. The mathematical expression for the transformation is as follows:








[



u




v




1



]

=


K
[



R


t



]

[




X
w






Y
w






Z
w





1



]


,






    • where R and t are the rotation matrix and translation vector of the camera with respect to the global coordinate system, which are known as extrinsic parameters of the camera. K is a matrix representing intrinsic parameters of the camera.









K
=


[




f

d
x




0



u
0





0



f

d
y





v
0





0


0


1



]

.





The camera projection matrix P is given by: P=K[R t].


The camera calibration is then a task of determining the camera projection matrix P given the global coordinates (X,Y,Z) of some 3D points and their corresponding image pixel coordinates (u,v), which are denoted as:







[




u
i






v
i





1



]

=

P
[




X
i






Y
i






Z
i





1



]





Each pair of 2D and 3D counterpart points provides two linear equations about the camera projection matrix P, that is, n pairs of 2D and 3D counterpart points provide 2n linear equations. When n is greater than or equal to 6, a direct linear transformation can be used to solve this problem. Therefore, the camera projection matrix can be obtained through capturing an image with at least 6 non-coplanar 2D and 3D counterpart points.


The processor 103 is configured to process the captured images and calculate the range of motion of the joint of the hand. More specifically, the processor 103 may be configured to: process the captured prior-movement images to obtain a plurality of prior-movement key point positions; process the captured post-movement images to obtain a plurality of post-movement key point positions; and calculate the range of motion of the joint based on the plurality of prior-movement key point positions and the plurality of post-movement key point positions.


Preferably, the machine vision-based system 100 further comprises a display 104 configured to demonstrate the hand movement to the subject.


Referring to FIG. 2 for the following description. In accordance with one embodiment of the present invention, a machine vision-based method S200 is provided for determining a range of motions of joints of a hand of a subject. The machine vision-based method S200 comprises:

    • S101: facilitating the subject to perform a hand movement designed for hand function assessment;
    • S102: capturing at least two prior-movement images of the hand when the hand has a neutral posture before performing the hand movement;
    • S103: capturing at least two post-movement images of the hand when the hand has an assessment posture after performing the hand movement;
    • S104: processing, using a neural network, the captured prior-movement images to obtain a plurality of prior-movement key point positions;
    • S105: processing, using the neural network, the captured post-movement images to obtain a plurality of post-movement key point positions;
    • S106: calculating the range of motion of the joint based on the plurality of prior-movement key point positions and the plurality of post-movement key point positions.


Referring to FIG. 3, in various embodiments, the plurality of prior-movement key point positions is obtained by mapping 2D positions of a plurality of sampling points of the hand in the prior-movement images into the projection 3D space through the camera matrixes to obtain a plurality of 3D prior-movement key point positions; and projecting the plurality of 3D prior-movement key point positions on a projection plane to obtain the plurality of prior-movement key point positions.


Similarly, the plurality of post-movement key point positions is obtained by mapping 2D positions of the plurality of sampling points of the hand in the post-movement images into the 3D space through the camera projection matrixes to obtain a plurality of 3D post-movement key point positions; and projecting the plurality of 3D post-movement key point positions on the projection plane to obtain the plurality of post-movement key point positions.


Referring to FIG. 4, the mapping of 2D positions from the prior-movement/post-movement images into the 3D space can be regarded as a triangulation problem referring to a reverse process of camera projection which involve reconstruction of the coordinates of a point X in the 3D space given the camera projection matrix and the pixel coordinates (u1,v1) and (u2,v2) of imaged points in the images captured by the corresponding cameras.


In practical computations, the projection lines of the two cameras might not exactly intersect. Therefore, for multi-view images captured by multiple cameras respectively, the 2D to 3D mapping can be formulated as an optimization problem of minimizing the reprojection error:





min Σin∥PiX−xi2,

    • where n represents number of view angles (or cameras), i denotes the i-th camera, Pi represents the projection matrix of the i-th camera, and xi=(ui,vi) represents the pixel coordinates of the imaged point in the i-th view angel, and X represents the desired coordinates of the object point.


In some embodiments, the hand movement is designed with respect to a kinematic model of the hand. By analyzing the anatomical model (as shown in FIG. 5A) of the hand, the kinematic model of the hand can be obtained. Referring to FIG. 5A, distal interphalangeal (DIP) joints 501 and proximal interphalangeal (PIP) joints 502 are collectively referred to as interphalangeal joints. Each of the four fingers has two interphalangeal joints, while the thumb has only one interphalangeal joint. The interphalangeal joints are hinge joints, allowing only flexion and extension movements with one degree of freedom (DOF).


The metacarpophalangeal (MCP) joint 503, also known as the knuckle joint, is a ball-and-socket joint, allowing flexion/extension and abduction/adduction movements with two DOF. However, in clinical assessments, based on the assumption that the abduction/adduction of the MCP joint of the thumb is associated with the abduction/adduction of the carpometacarpal (CMC) joint 504 of the thumb, only flexion and extension evaluations (one DOF) are usually performed for the MCP joint of the fingers.


The CMC joint 504 is the carpometacarpal joint of the thumb, also known as TM joint, and it is a saddle joint, enabling flexion, extension, abduction, adduction, and circumduction movements with two DOF. Therefore, in the present invention, a 26 DOF kinematic model for the hand as shown in FIG. 5B is employed.



FIGS. 6A to 6G illustrate seven hand movements respectively designed for hand function assessment.


Movement A (as shown in FIG. 6A) requires the subject to first extend all five fingers and then forcefully flex them to the maximum extent, is designed for demonstrating the flexion/extension range of motion of the nine interphalangeal joints in the hand.


Movement B (as shown in FIG. 6B) requires the subject to keep the distal interphalangeal joints of the four fingers extended to form a straight first and then touch the fingertips as far as possible to the wrist. Movement B displays the flexion/extension range of motion of the four metacarpophalangeal joints in the hand.


Movements C and D (as shown in FIGS. 6C and 6D respectively) involve the abduction and adduction of the four metacarpophalangeal joints respectively. The subject is required to fully spread the fingers apart and then bring them together to the maximum extent, the range of motion for finger abduction/adduction can then be assessed.


Movements E and F (as shown in FIGS. 6E and 6F respectively) involve abduction and adduction of the thumb carpometacarpal joint, including palmar abduction (E) and radial abduction (F).


Movement G (as shown in FIG. 6G) requires the subject to relax the four fingers while fully flexing the thumb to the maximum extent. This movement demonstrates the maximum flexion of the thumb interphalangeal joint and metacarpophalangeal joint.


These specific movements are designed to assess the various ranges of motion of different joints in the hand, allowing therapists to evaluate the hand's flexibility, stability, and functionality of each joint in the hand. This assessment is common in clinical practice, especially for patients undergoing hand rehabilitation after injuries, hand diseases, or hand surgeries. Such assessments can also be used in research involving patients with hand diseases or movement disorders to understand how different conditions affect the range of motion in hand joints, providing a more scientific basis for treatment and rehabilitation strategies.


Referring to FIG. 7, depending on the declination angles of joints to be measured in determining the range of movement, the projection plane may be a plane H in parallel with a palm plane of the hand and/or a plane V perpendicular to the palm plane. The projected prior-movement key point positions and post-movement key point positions are then used to calculate the ROM.


For example, for movements A, B and E, the plurality of 3D post-movement key point positions is projected on the plane perpendicular to the palm plane to obtain the plurality of post-movement key point positions for calculating the ROM. For movements C, D, F and G, the plurality of 3D post-movement key point positions is projected on the plane in parallel with the palm plane to obtain the plurality of post-movement key point positions for calculating the ROM.


The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications that are suited to the particular use contemplated. While the methods disclosed herein have been described with reference to particular operations performed in a particular order, it will be understood that these operations may be combined, sub-divided, or re-ordered to form an equivalent method without departing from the teachings of the present disclosure. Accordingly, unless specifically indicated herein, the order and grouping of the operations are not limitations. While the apparatuses disclosed herein have been described with reference to particular structures, shapes, materials, composition of matter and relationships . . . etc., these descriptions and illustrations are not limiting. Modifications may be made to adapt a particular situation to the objective, spirit and scope of the present disclosure. All such modifications are intended to be within the scope of the claims appended hereto.

Claims
  • 1. A machine vision-based method for determining a range of motion of a joint of a hand of a subject, comprising: facilitating the subject to perform a hand movement at a preset position;capturing at least two prior-movement images of the hand when the hand is in a neutral posture before performing the hand movement;capturing at least two post-movement images of the hand when the hand is in an assessment posture after performing the hand movement;processing the captured prior-movement images to obtain a plurality of prior-movement key point positions;processing the captured post-movement images to obtain a plurality of post-movement key point positions; andcalculating the range of motion of the joint based on the plurality of prior-movement key point positions and the plurality of post-movement key point positions.
  • 2. The machine vision-based method of claim 1, further comprising calibrating each of at least two cameras, which are configured to capture images of the hand from at least two perspectives respectively, with respect to a 3D space to obtain a respective camera projection matrix.
  • 3. The machine vision-based method of claim 2, wherein the plurality of prior-movement key point positions is obtained by: mapping 2D positions of a plurality of sampling points of the hand in the prior-movement images into the 3D space through the camera projection matrixes to obtain a plurality of 3D prior-movement key point positions; andprojecting the plurality of 3D prior-movement key point positions on a projection plane to obtain the plurality of prior-movement key point positions; andthe plurality of post-movement key point positions is obtained by: mapping 2D positions of the plurality of sampling points of the hand in the post-movement images into the 3D space through the camera projection matrixes to obtain a plurality of 3D post-movement key point positions; andprojecting the plurality of 3D post-movement key point positions on the projection plane to obtain the plurality of post-movement key point positions.
  • 4. The machine vision-based method of claim 1, wherein the projection plane is a plane in parallel with a palm plane of the hand or a plane perpendicular to the palm plane.
  • 5. The machine vision-based method of claim 1, wherein the plurality of sampling points of the hand include: finger tips, distal interphalangeal (DIP) joints, proximal interphalangeal (PIP) joints and metacarpophalangeal (MCP) joints of the hand.
  • 6. The machine vision-based method of claim 5, wherein the hand movement is designed by employing a kinematic model presuming a plurality of degrees of freedom of the hand.
  • 7. The machine vision-based method of claim 6, wherein the plurality of presumed degrees of freedom of the hand includes: flexion and extension movements of the distal interphalangeal (DIP) joints and the proximal interphalangeal (PIP) joints; andflexion, extension, abduction, adduction, and circumduction movements of the metacarpophalangeal (MCP) joints.
  • 8. A machine vision-based system for determining a range of motion of a joint of a hand of a subject, the system comprising: a supporter configured to facilitating the subject to perform a hand movement at a preset position;at least two cameras configured to: capture at least two prior-movement images of the hand when the hand is in a neutral posture before performing the hand movement; andcapture at least two post-movement images of the hand when the hand is in an assessment posture after performing the hand movement; anda processor configured to: process the captured prior-movement images to obtain a plurality of prior-movement key point positions;process the captured post-movement images to obtain a plurality of post-movement key point positions; andcalculate the range of motion of the joint based on the plurality of prior-movement key point positions and the plurality of post-movement key point positions.
  • 9. The machine vision-based system of claim 8, wherein each of the at least two cameras are calibrated with respect to a 3D space to obtain a respective camera projection matrix and configured to capture images of the hand from a corresponding perspective.
  • 10. The machine vision-based system of claim 9, wherein the processor is further configured to obtain the plurality of prior-movement key point positions by: mapping 2D positions of a plurality of sampling points of the hand in the prior-movement images into the 3D space through the camera projection matrixes to obtain a plurality of 3D prior-movement key point positions; andprojecting the plurality of 3D prior-movement key point positions on a projection plane to obtain the plurality of prior-movement key point positions; andthe processor is further configured to obtain the plurality of post-movement key point positions by: mapping 2D positions of the plurality of sampling points of the hand in the post-movement images into the 3D space through the camera projection matrixes to obtain a plurality of 3D post-movement key point positions; andprojecting the plurality of 3D post-movement key point positions on the projection plane to obtain the plurality of post-movement key point positions.
  • 11. The machine vision-based system of claim 8, wherein the projection plane is a plane in parallel with a palm plane of the hand or a plane perpendicular to the palm plane.
  • 12. The machine vision-based system of claim 8, wherein the plurality of sampling points of the hand include: finger tips, distal interphalangeal (DIP) joints, proximal interphalangeal (PIP) joints and metacarpophalangeal (MCP) joints of the hand.
  • 13. The machine vision-based system of claim 8, wherein the hand movement is designed by employing a kinematic model presuming a plurality of degrees of freedom of the hand.
  • 14. The machine vision-based system of claim 13, wherein the plurality of presumed degrees of freedom of the hand includes: flexion and extension movements of the distal interphalangeal (DIP) joints and the proximal interphalangeal (PIP) joints; andflexion, extension, abduction, adduction, and circumduction movements of the metacarpophalangeal (MCP) joints.
  • 15. A non-transitory computer-readable storage medium storing a program including instructions for performing the machine vision-based method of claim 1.
  • 16. The non-transitory computer-readable storage medium of claim 15, wherein the method further comprises: calibrating each of at least two cameras, which are configured to capture images of the hand from at least two perspectives respectively, with respect to a 3D space to obtain a respective camera projection matrix.
  • 17. The non-transitory computer-readable storage medium of claim 16, wherein the plurality of prior-movement key point positions is obtained by: mapping 2D positions of a plurality of sampling points of the hand in the prior-movement images into the 3D space through the camera projection matrixes to obtain a plurality of 3D prior-movement key point positions; andprojecting the plurality of 3D prior-movement key point positions on a projection plane to obtain the plurality of prior-movement key point positions; andthe plurality of post-movement key point positions is obtained by: mapping 2D positions of the plurality of sampling points of the hand in the post-movement images into the 3D space through the camera projection matrixes to obtain a plurality of 3D post-movement key point positions; andprojecting the plurality of 3D post-movement key point positions on the projection plane to obtain the plurality of post-movement key point positions.
  • 18. The non-transitory computer-readable storage medium of claim 15, wherein the projection plane is a plane in parallel with a palm plane of the hand or a plane perpendicular to the palm plane.
  • 19. The non-transitory computer-readable storage medium of claim 15, wherein the plurality of sampling points of the hand include: finger tips, distal interphalangeal (DIP) joints, proximal interphalangeal (PIP) joints and metacarpophalangeal (MCP) joints of the hand.
  • 20. The non-transitory computer-readable storage medium of claim 15, wherein the hand movement is designed by employing a kinematic model presuming a plurality of degrees of freedom of the hand.
  • 21. The non-transitory computer-readable storage medium of claim 20, wherein the plurality of presumed degrees of freedom of the hand includes: flexion and extension movements of the distal interphalangeal (DIP) joints and the proximal interphalangeal (PIP) joints; andflexion, extension, abduction, adduction, and circumduction movements of the metacarpophalangeal (MCP) joints.