The present invention relates to a device and method which is configured to allow a user to specify goal and path of states to that goal in a consistent manner via biological signals such as those derived from brain or eye-tracker recordings. One exemplary application of the present invention is for a treatment and rehabilitation of patients with impairment of motor function.
Machines that enable the user to express intended control signals can be broadly referred to as bio-prosthetics. There has been interest in developing bio-prostheses that can circumvent user's inability to reliably activate specific nerves or muscles, such as for patients with spinal cord lesions, stroke, tremor, or myopathies. Indeed, reaching movements to a ball with a robot arm or active brace device, or moving an on-screen mouse to an icon, are some examples for which the user expresses control signals through a bio-prosthetic that specify a goal and a trajectory of states of the prosthetic to achieve that goal.
A control bio-prosthetic generally can involve the mapping of user-derived signals, from brain, eye, muscle, or otherwise, onto control signals that specify aspects of the target state and the path to that target state. One of the problems addressed by the exemplary embodiments of the present invention is the problem of combining user-derived signals relating to target as well as path into one consistent set of control signals that can be used in real-time for moving the bio-prosthetic from the current state, through the desired path, to the desired target.
Currently available real-time methods either allow the user to specify the path to the goal (as described in Wu, W. et al., “Modeling and decoding motor cortical activity using a switching Kalman filter,” IEEE Trans. On Biomedical Engineering 51 (6), pp. 933-942 Jun. 2004), or the goal itself (as described in Musallam, S., et al., “Cognitive Control Signals for Neural Prosthetics.” Science, 305(5681), pp. 258-262). The result in the first case is that the path is unconstrained by the goal, and in the second case is that there is no real-time user-controlled ability to specify the path to the goal.
One conventional method described in C. Kemere, et al. “Model-based neural decoding of reaching movements: a maximum likelihood approach,” IEEE Trans. On Biomedical Engineering, Jun. 2004, pp. 925-932, and C. Kemere et al., “Model-Based Decoding of Reaching Movements for Prosthetic Systems,” Proc. Of the 26th th Annual Conf. Of the IEEE EMBS, Sep. 2004, pp. 4524-4528, in the specific context of decoding reaching arm movements, attempted to combine hand target and current kinematic information from motor cortical-derived signals in order to estimate trajectory states. However, this method uses a predefined template trajectory distribution or a discrete database of arm movements to be pre-recorded for each potential target location. The brain activity is matched to one of these template arm motions to select a target and the current arm position according to the best template arm movement. Moreover, this method uses batch-mode processing of all data, and may require a numerical maximization procedure.
In contrast, the free arm movement estimation literature has largely relied on either no models or random walk-type models, both for computational simplicity and for robustness in the face of uncertain movement constraints as described in A.B. Schwartz, “Cortical Neural Prosthetics”, Annu. Rev. Neurosci., vol. 27, pp. 487-507, Mar. 2004.
Certain exemplary embodiments of the present invention provide a system, method, an executable arrangement, and computer-accessible medium for providing state information relating to a simulated or actual execution of one or more trajectories of an object. A first set of data can be obtained prior to execution of a trajectory, which can include information relating to a portion of the trajectory including, e.g., an estimation of a particular state of the trajectory at a later point in time. A second set of data can then be obtained during execution of the trajectory, which can include information relating to a portion of the trajectory, e.g., the time at which the object reaches one particular state of the trajectory. State information may then be generated by associating the first and second data sets, independently of data associated with predetermined trajectories. The state information may be generated in real time and/or in a recursive manner, and may optionally be used to affect the execution of the object trajectory. The data sets may be associated with signals obtained from one or more anatomical structures, such as a brain, an eye, a muscle, a tongue, and the like.
In further exemplary embodiments of the present invention, the first and/or second data set may exclude a predetermined arrival time for a particular state.
In further exemplary embodiments of the present invention, a third data set associated with a minimal probabilistic constraint on the trajectory and the current position of the object may also be used in generating the state information relating to the trajectory.
According to exemplary embodiments of the present invention, a generic model of free arm movement can be combined with information about the target state to produce a generic model for reaching arm movement. Uncorrelated increments may also facilitate the use with real-time recursive estimation procedures, such as Kalman filters.
The exemplary embodiments according to the present invention can utilize such models, which may represent a set of prior states for any method of estimating arm movements, including point process filters, Kalman filter variants, particle filters, or general probabilistic inference. Measurements from any device or brain region can be incorporated into this estimation procedure, including local field potentials and spiking activity.
According to one exemplary embodiment of the present invention, continuous time surveillance methods as described in D.A. Castanon et al. ¢Algorithms for the incorporation of predictive information in surveillance theory”, Int. J Systems Sci., vol. 16, no. 3, pp. 367-382, 1985 can be adapted for discrete time. The exemplary derivations described herein follow an approach similar to a discrete-time backwards Markov model construction as provided in G. Verghese et al., “A Further Note on Backwards Markovian Models”, IEEE Trans. Info. Theory, vol. IT-25, no. 1, Jan. 1979.
Further, a goal-directed reach state equation can be utilized by the exemplary embodiments of the present invention. In addition, an augmented state space may be used to accommodate concurrent target dynamics. These exemplary methods (and systems implementing such methods) can be used to estimate arm movement from simulated cortical activity during a reach.
For example, exemplary embodiments of the method, system, executable arrangement, and computer-accessible medium according to the present invention can provide for a maximum flexibility in allowing the user to specify a target and a continuous path to that target. Such exemplary embodiments may be compatible with real-time processing of signals. In contrast to the exemplary embodiments of the present invention, some conventional methods and devices require the use of previously provided and static trajectory specifying information. In addition, a number of conventional methods and devices require the use of a pre-recorded database of movements, a control signal of specific dimension, or probabilistic template matching to a discrete database of pre-recorded arm movements.
For example, the exemplary embodiments of the present invention can be compatible with real-time signal processing methods. For example, according to one exemplary embodiment of the present invention, and in contrast with conventional techniques, batch processing is not necessary and control signals can be updated efficiently with the last observation of the biological signals. In addition, according to another exemplary embodiment of the present invention, a number of calculations are precomputed, again facilitating real-time application. Further, according to another exemplary embodiment of the present invention, no numerical optimization is needed and the optimal solution can be analytically expressed. Indeed, these benefits represent improvements, and can facilitate a facile production of goal-directed control signals.
For example, the exemplary embodiments of the present invention can facilitate any interface between a person and a machine where the person can describe to the machine an intended target position and path to the target with any level of certainty. One exemplary application of such exemplary embodiment can be a bio-prosthetic application, where brain or eye-tracker derived signals are used to communicate a user's intent in controlling a machine. Further, patients with sensorimotor deficits (including central and peripheral neuropathies, as well as muscle disorders) may benefit from the bio-prosthetics that allow them to express control signals without the direct activation of affected muscles or nerves. Goal directed movements may be common to our everyday function. The exemplary embodiments of the present invention can provide a flexible, real-time solution to facilitating such classes of movements, and can seamlessly integrates with the expression of continuous movements.
These and other objects, features and advantages of the present invention will become apparent upon reading the following detailed description of embodiments of the invention, when taken in conjunction with the appended claims.
Further objects, features and advantages of the invention will become apparent from the following detailed description taken in conjunction with the accompanying figures showing illustrative embodiments of the invention, in which:
The exemplary embodiments of the device and system according to the present invention can be implemented using the following exemplary techniques.
I. Incorporation of Target Information
Initially, any P-dimensional linear time-varying discrete-time state space model of free arm movement can be provided. At time t, the vector xt describes the arm state, At is an invertible state transition matrix, and wt may be a zero mean Gaussian increment:
xt=Atxt−1+wt (1)
E[wtw′T]=Qtδt−T (2)
The initial condition can be specified as
x0˜N(m0,π0) (3)
An observation of target position specified with uncertainty for known arrival time T can be written as follows:
yT=xT+vT (4)
where vt˜N(0,πT) denotes the observation noise. The Bayes least squares estimate of each increment wt conditional on the noisy observation YT of the target location and the present location xt−1 are to be obtained. This can be equivalent to the linear least squares estimate (“LLSE”) for jointly Gaussian distributions. The standard LLSE estimate of wt from YT can be provided as follows:
{circumflex over (w)}t(yT)=E[wt]+cov(wt,yT)cov−1(yT,yT)(yT−E[yT]) (5)
The further increment can correspond to an error of the LLSE estimate, with covariance given by:
cov(et)=cov(wt)−cov(wt,yT)cov−1(yT)cov′(wt,yT) (6)
The expected value of wt may be zero. To derive expressions for the expected value and covariances involving yT, yT is provided in terms of xt−1,VT and the intervening increments wi for t≦i≦T; that is
where ∅(t, s) denotes the state transition matrix that maps the state xs to the state xt,
It follows that the expected value of YT is
E[yT]=φ(T,t−1)x(t−1) (8a)
with covariance terms
It should be understood that in the formulas below, φT should be considered approximately or substantially the same as φ′. Then, the mean and covariances of YT are determined and substituted into the LLSE estimation formula (5) above to obtain the mean of the increment based on target information,
The error covariance can be derived from formula (6) above and simplified to obtain
Therefore, the increment wt together with a provided xt−1 and noisy target observation YT may have a mean {circumflex over (w)}t and covariance cov(Ct) as determined above.
For example, the quantity Π(t, T) can be determined recursively starting at Π(T, T). To obtain this result, it is possible to compare the equations for Π(t−1, T) and Π(t, T ). The consequent recursion can be written as
For an estimation, Π(t, T) can be determined with each new observation, intermittently, or not at all.
To complete the equivalent reach state equation, the initial state and covariance may be updated with standard LLSE formulas,
ΠS=(Π−10+Π−1(0,T))−1 (15)
xs=ΠS(0)(Π−10m0+Π−1(0,T)φ(0,T)yT) (16)
In summary, the equivalent state equation for reaching movements may be provided by
xt=Atxt−1+ut+et (17)
ut={circumflex over (w)}t(yT) (18)
et˜N(0,cov(et)) (19)
x0˜N(xs,Πs) (20)
The exemplary embodiment of the process resulting from the new state equation (17) may also be Markov, because xt is a function of the state at only the preceding time step xt−1 . Additionally, based on the orthogonality principle, the C, can be uncorrelated.
An exemplary flow diagram of the method 100 according to the present invention is shown in
The above equation (17) provides an intuition about the evolution of the arm trajectory under the exemplary reaching model according to the present invention. As the hand moves closer to the target in time, the trajectory can become more constrained by the target location. This can be accomplished by a combination of the forcing term u, and the noise term C,. With time, the forcing term more insistently pushes the arm to a path that sets it on course for the target. The covariance in the noise term can taper in proportion to target uncertainty as it becomes more apparent that specific changes in state are needed to bring the arm to the target at time T.
II. Uncertain Reach Duration
Typically, at times, it is may not be possible to confirm that an arm will meet the target at precisely time T. More realistically, it may be possible to specify a distribution of arrival times that are observed. It is possible to express an uncertainty in arrival time as a distribution p(T) on T. In this context, the results from the previous section for fixed arrival time can be interpreted as a state evolution equation conditioned on T. It is possible to employ Bayes rule to describe the distribution of increments that results from uncertain arrival time, as
As described above, p(wt\xt−1, xT, T)˜N(ŵt, cov(ŵt)). However, this distribution is not Gaussian in the T parameter. Consequently, even if the p(T) distribution was Gaussian, the resulting distribution p(wt\xt−1, xT) would likely not be Gaussian. While non-Gaussian increments are generally incompatible with standard LLSE techniques, computational approaches are available to propagate non-Gaussian distributions through state equations, including particle filters, mixture of Gaussians, and numerical integration of the above Bayes rule formula.
In a numerical integration, a discrete set of M arrival times can be considered, and the increment estimates for each arrival time may be combined by weighting their respective probabilities,
The resulting density may be approximated with a Gaussian density to allow compatibility with LLSE.
An alternative exemplary embodiment of the method according to the present invention can execute filters in parallel that assume different arrival times. The estimated state at each point may be an average of the individual filter predictions, weighted by the probability of the assumed target time. The actual target time probabilities can be adjusted based on the difference between the predicted and observed measurement, which may be referred to as an innovation in Kalman filters. The innovation for each filter can be obtained from a zero mean Gaussian with variance of the predicted observation, substantially equal in the Kalman filter to the state prediction variance sent through the observation equation.
With a data likelihood provided for each filter, the probability of each arrival time can be updated by Bayes rule, multiplying each data likelihood by the old probability of arrival time and normalizing. This exemplary update of probabilities may provide a dynamical estimate of target arrival time. Various forms of this exemplary method are described in the publications referenced herein.
III. Updating Target Estimate
Observations regarding the reach path can also be used to refine estimates of the target if path and target are dependent. To support a recursive estimation of target position, it is possible to augment the state space with XT to include target position and target velocity variables. The resulting state equation can be provided as
where
Ψ=At−Qtφ′(t−1,t)Π−1(t−1,T) (22)
T=Qtφ′(t−1,t)Π−1(t−1,T)φ(t−1,T) (23)
and the increment can be
Et=[et 0 ]′ (24)
Uncertainty can be added to this increment to track possible drifts in target states.
Such a one-state equation can support a real-time decoding of concurrent or sequential neural activity likely relating in various degrees to the path and target of the reach. When only trajectory information is available, the model can be reduced to a free-arm movement, where ut is zero and C, retains approximately the same statistics as wt. When only the target information is available, the model dynamics can travel to the target in proportion to the uncertainty about the target based on ut. Because the target position and velocity become state variables, refined estimates of the target can be available even as the reach proceeds.
IV. Exemplary Applications
To illustrate the flexibility of the above reach state equation, it is possible to simulated neural spiking data from primary motor cortex in response to a simulated two dimensional trajectory. A linear update point process filter based on the probabilistic principles of Kalman filters can be used to reconstruct the trajectory using a free arm movement state equation and the reach state equation. The point process filtering method has been described in detail in U.T. Eden, “Dynamic Analysis of Neural Encoding by Point Process Adaptive Filtering”, Neural Computation, vol. 16, no. 1, 2004. Another article, i.e., Alan S. Willsky, 6.433 Recursive Estimation: Supplementary Notes, MIT Course 6.433 Notes, Topic 5.4, p. 18, 1994, may also be relevant thereto.
The general point process observation equation for the ith neuron at time t can be approximated as follows
where Hit denotes the history of state xt and spikes ΔNit per time interval δ up to (but possibly excluding) timestep t. The conditional intensity function described and shown above can follow a primary motor neuron cosine tuning technique as described in D. Moran et al., “Motor cortical representation of speed and direction during reaching,” J Neurophys., vol. 82, no. 5, pp. 2676-2692, 1999, and as provided by the following formula:
λ(t\vx,vy)=exp(β1+β2 vx+β3vy) (26)
The performance in decoding reaching movements using the free arm movement model (a) can be compared against results using the reach state equation (17) provided herein. Exemplary results thereof are illustrated for one example in
In particular,
The exemplary embodiment of the point process filter results may track the actual trajectory more closely with the reach state equation than with the free movement state equation. The reach state equation can be used to generate trajectories, from which a spiking activity may be simulated with a receptive field model of primary motor cortex. Point process filter reconstructions using a free movement state equation and a reach state equation can be compared against true movement values. In these examples shown in
Decoding under a simple model violation can then be examined. For example, as shown in
An exemplary refinement of target estimates from the trajectory related neural activity may then be illustrated. Because the target state augmented reach model of equation (21) can be employed, target estimates can be refined with each sample of neural data observed during the movement. Target estimation with the augmented state equation may be determined for one trial. The initial estimate of the target can be intentionally set to be incorrect at (1 m, 1 m) and with a variance of I m2 that is large relative to the distance to the true target location at (0.25 m, 0.25 m). Subsequent target estimates may be produced using a simulated neural spiking activity that can relate directly to the path rather than the target.
An exemplary decoding performance for the target estimation trial is illustrated in
Further, the mean squared errors for trajectory reconstruction may be examined as the target state became more uncertain. One common simulated set of neural data may be used to make a performance comparison between the two methods. Mean squared errors may be averaged over 10 trials for the point process filter using the free and reach state equations separately. The exemplary results are shown in
The individual performance of the exemplary models may be specific to the set of neuron tuning curves that is selected. Additionally, the relative improvement observed in
For example,
The results from other examples of the reconstructions are shown in
An exemplary embodiment of a system for implementing the present invention is shown in
The external information 940 and information contained in the database 950 can be communicated to computer 920. The computer 920 may be configured to calculate state information for the bio-prosthetic 970 based at least in part on these informations in accordance with the exemplary embodiment of the methods of the present invention described above. State information determined by computer 920 may be communicated to a display 960. The display 960 may comprise any suitable display device, including but not limited to a video monitor, a printer, a data storage medium, a sound generator, and the like. The computer 920 may also communicate state information to a controller 930, which can be in communication with, or optionally incorporated into, the bio-prosthetic 970. The controller 930 can be capable of providing signals to the bio-prosthetic 970 to at least partially direct movement of the bio-prosthetic 970 based on the state information provided. The computer 920 can include a hard drive, CD ROM, RAM, and/or other storage devices or media which can include thereon software, which can be configured to execute the exemplary embodiments of the method of the present invention.
V. Conclusion
The exemplary reach state equation, and the exemplary system and method which can use such equations, can represent the Bayesian-optimal extension of the general autoregressive state equation that has been popular in the decoding of the free arm movements with Kalman and point process filters. The resulting exemplary model may provide a minimum constraint placed on a free arm movement based on observation of its target state.
The exemplary model may be a prior on the reach states, and can be used in conjunction with the estimation procedures that employ any modality, such as local field potentials or spikes, or any formulation between the measured signals and desired reaching movement. The result can be extended to accommodate uncertain reach duration.
One preferable exemplary feature of the model of the method and system of the exemplary embodiment of the present invention is that by switching between the informative and uninformative target estimates, the reconstruction can alternate between reaching and free arm movements. For neural prosthetics, preserving the dynamics of the exemplary system between reaching and free movement modes may be easier for the user than switching to an entirely different set of priors between these two movement modes.
Because the reach state equation can place small constraints on the motion, it is likely to be more robust to model violations than trajectory models derived from optimization of specialized cost functions as described in E. Todorov, “Optimality principles in sensorimotor control”, Nature Neurosci., vol. 7, no. 9, pp. 907-915, Sept. 2004, or models based on empirical databases described in C. Kemere et al., Model-Based Decoding of Reaching Movements for Prosthetic Systems“, Proc. 26th Annu. Int. Conf IEEE EMBS, San Francisco, Calif., pp.4524-4528, Sept. 1-5, 2004.. This benefit may be important in providing the user both flexibility and control. The method and device according to the exemplary embodiment of the present invention may also provide a closed form solution and may be more efficient and easy to implement than conventional methods requiring numerical optimization.
The exemplary reach state equation may be a basic example of a closed loop stochastic control model in that the forcing term is determined by the current and end states rather than a preformed control sequence. Other exemplary approaches in accordance with the present invention may examine the use of Monte Carlo methods with nonlinear system dynamics or stochastic optimal control with relatively unconstrained cost functions such as minimum target variance.
The foregoing merely illustrates the principles of the invention. Various modifications and alterations to the described embodiments will be apparent to those skilled in the art in view of the teachings herein. It will thus be appreciated that those skilled in the art will be able to devise numerous systems, arrangements and methods which, although not explicitly shown or described herein, embody the principles of the invention and are thus within the spirit and scope of the present invention. In addition, all publications referenced above are incorporated herein by reference in their entireties.
This application claims priority from U.S. Patent Application No. 60/647,590, filed Jan. 26, 2005, the entire disclosure of which is incorporated herein by reference.
The research leading to the present invention was supported, at least in part, by National Institute of General Medical Sciences, “Medical Scientists Training Program”, Grant number 5T32GM07753-26 and National Institute on Drug Abuse, “Dynamic Signal Processing Analysis of Neural Plasticity”, Grant number R01 DA015644. Thus, the U.S. government may have certain rights in the invention.
Number | Date | Country | |
---|---|---|---|
60647590 | Jan 2005 | US |