This invention relates generally to automated image recognition and, in particular, to computer-based behavior recognition methods and apparatus.
Differentiating between normal human activity and suspicious behavior is a difficult task, whether performed by a sensing device or a human observer. A human observer would find such a task tedious and costly to perform in money and time. Fortunately, a sensing system is not bothered by ‘tedious’ tasks. Such a system could be implemented to prune out obviously normal behavior, and tag human activities which could be suspicious and would therefore need more attention by a human operator.
However, such “behavior recognition” systems have not been developed due to the difficulty of identifying and classifying such motions. Consider an urban environment as illustrated in
The present invention is directed to the automated classification and identification of human activities. The requisite heuristics involve recognition of information-bearing features in the environment, and the determination of how those features relate to each other over time. The approach resides in a behavior recognition sensor system which identifies simultaneously created gestures using a gesture recognition system. The underlying gesture recognition system performs the task of determining the behavior (state) of objects in motion.
According to the invention, general kinematic relationships for various bodies (human, animal, robotic) are modeled. Next, specific link combinations for each are parameterized and modeled. This enables the system to recognize such motions as the various human motion gaits (if the links are legs), the throwing of an object (if the links are arms), or any other type of motion and orientation. A whole-body (human, vehicle, or other) link representation and dynamic model is then developed.
The state of an object (a human or a device) can take three forms.
Whenever an object is identified, the system preferably first determines which state the object is in based upon gesture recognition. The term “gesture” not only refers to the (dynamic) motion of an object, but also to the state of an object that is not moving (static). For example, a gun being pointed by a human should definitely be classified as a gesture. Therefore, determining behaviors also means identifying combinations of static and dynamic gestures.
An object is composed of many connected non-deformable links. For example, a person is made up of a torso (one link), a head (one link), two arms (each with two links), two hands (one link each, not counting the fingers), two legs (each with two links), and two feet (one link each). Each link of an object has its own gesture motion, which, when examined in relation to the other links in the object, can be used to determine the overall state of the system. For example, when just examining the gestures created by the foot and knee joint of a human, one can see that the motions of those features are different depending on whether a person is walking or running. Even when not in motion, the relationship of those non-deformable links gives rise to information about the object's state. A person who is standing still will have a different kinematic link relationship to one who is sitting, or lying down.
Such gross human torso features can be readily identified (using a model that matches general body features) with the vision system described in co-owned U.S. Pat. No. 6,681,031. Building on this technology, the gesture recognition module of this invention determines if a dynamic motion is occurring, and uses that information with kinematic link relationships, develop a hypothesis about the overall state of the object or objects in the field of view.
In a sense, then, the invention provides an automatic method for classifying and categorizing such dynamic motions and gaits. Such state recognition is not limited to humans (and other animals), however. A vehicle's state can also be determined by examining the various moving parts, such as the body and the tire motions. Even unknown devices (such as mobile robots) can be classified by examining their motion and behavioral characteristics.
Although subject visualization is preferred, alternative approaches to gait/behavior recognition may be employed according to the invention, including electro-mechanical methods of tracking the body to identify human motion. This includes using:
The core of the behavior recognition system is a set of dynamic and static gesture recognition modules. This section details the overall gesture recognition system used by the behavior recognition system.
The Gesture Recognition System
The specific components of the gesture recognition system is detailed in
G: Gesture Generation
S: Sensing (vision)
I: Identification Module
T: Transformation
R: Response
At a high level, the flow of the system is as follows. Within the field of view of one or more standard video cameras, a gesture is made by a person or device. During the gesture making process, a video capture card is capturing images, producing image data along with timing information. As the image data is produced, they are run through a feature tracking algorithm which outputs position and time information. This position information is processed by static and dynamic gesture recognition algorithms. When the gesture is recognized, a command message corresponding to that gesture type is sent to the device to be controlled, which then performs and appropriate response. The five modules are detailed below.
Gesture Creator
In the Gesture Creator module, a human or devices creates a spatial motion to be recognized by the sensor module. If one camera is used, then the motion generated is two dimensional and parallel to the image plane of the monocular vision system. For three dimensional tracking (as is also done with this system), stereo vision using two or more cameras are used.
This gesture recognition system is designed to recognize consistent yet non-perfect motion gestures and non-moving static gestures. Therefore, a human can create such gestures, as well as an actuated mechanism which could repeatedly create perfect gestures. Human gestures are more difficult to recognize due to the wide range of motions that humans recognize as the same gesture. We designed our gesture recognition system to recognize simple Lissagous gesture motions (repeating circles and lines), repeated complex motions (such as “come here” and “go away quickly” back and forth hand motions which we define as “skew” gestures), and static hand symbols (such as “thumbs-up”).
With regards to human generated gestures used for communication or device control, we chose gestures to be identified based on the following:
Humans should be able to make the gestures easily.
The gestures should be easily represented mathematically.
The lexicon should match useful gestures found in real world environments.
For the dynamic (circular and skew) gestures, these consist of one-dimensional oscillations, performed simultaneously in two or three dimensions. A circle is such a motion, created by combining repeating motions in two dimensions that have the same magnitude and frequency of oscillation, but with the individual motions ninety degrees out of phase. A “diagonal” line is another such motion. We have defined three distinct circular gestures in terms of their frequency rates: slow, medium, and fast. An example set of such gestures is shown in
The dynamic gestures are represented by a second order equation, one for each axis:
{dot over (x)}1=x2
{dot over (x)}2=θ1x1+θ2
More complex second order models are used to recognize more complex gestures (discussed later). This gesture model has no “size” parameter. θ1 is a frequency measure, and θ2 is a drift component. The gestures were named “large”, “small”, “fast”, and “slow” due to the human motions used to determine the parameters (see
For example, a total of 24 gestures are possible when the following are distinct gestures: clockwise and counter-clockwise circles, diagonal lines, one dimensional lines, and small and large circles and lines. Geometric constraints are required to expand the lexicon, because different gestures can result in the same parameters.
As with the previous gestures, additional gestures can be created from these two gestures by varying the phase relationships.
Phase relationships are determined as follows. During the gesture, the x's and y's (and z's, if the system is set up for three dimensions) minimum and maximum image plane positions are computed. If the x and y motions are out of phase, as in a circle, then when x or y is minimum or maximum, the other axis's velocity is large. The direction (clockwiseness in two dimensions) of the motion is determined by looking at the sign of this velocity component. Similarly, if the x and y motion are in phase, then at these extremum points both velocities are small.
Example dynamic gestures used for real world situations are taken from a 1987 Army Training Manual. A “Slow Down” gesture is a small x-line created to one side of the body (
Static gestures are represented as geometric templates. Four gestures are shown and are representative of the static gestures which can be represented and identified by this gesture recognition system. Additionally, language gestures, such as American Sign Language gestures, can also be recognized.
The example static gestures are:
The gesture recognition system identifies a moving gesture by its dynamics—that is, the structure of its positions in space over time. The system translates the motion information into parameters which are used to develop commands for controlling data outputs and actuated mechanisms. For example, the speed at which a person waves a robot away might directly affect a robot arm's velocity or a mobile robot's speed. In order for recognition to occur, a representation for human gestures is required, from which a computational method for determining and recognizing specific gestures can be derived.
Although we make these gestures in two and three dimensions, the explanation now detailed is described simply dimension as a basic one dimensional gesture as a simple example to clarify the distinction between the “shape” and the “dynamics” of a gesture. The techniques for identifying this basic gesture will be used to identify similar oscillatory motions occurring in two and three dimensions.
First, a dynamic system gesture representation is determined, both the model for representing the oscillatory gestures and parameter determination scheme was developed. For this system a Linear Least Squares method was an on-line computationally efficient technique which allowed us to use a linear-in-parameters gesture model.
The representative planar gesture used throughout this section to exemplify our method consists of a family of oscillating motions which form a (roughly) horizontal line segment (“x-line motion”). As discussed earlier, a human is incapable of reliably generating a perfect sinusoidal motion.
For this dynamic gesture recognition module, a computationally effective mathematical representation for the gesture plotted in
x(t)=?,
where “?” would be replaced with some structure based on measurable features which are used to classify the gesture. Of course, there are an infinite number of possible measurable features.
We can make the number of classifications (the “feature space” dimension) finite by restricting the form of the representations. Instead of representing gestures as x(t), the representation might be constrained through the use of a parameter vector, resulting in x(t,p). The feature space dimension is then equivalent to the number of parameters we store. For example, when:
x(t,p)=p0+p1t, t>0,
the only possible gestures that we can represent are lines described by the two parameters slope, p1, and intercept p0 (see
Even with a finite dimensional representation, each unique motion is represented by its own distinct parameters. However, our intuition about human gestures tells us that certain distinct motions should have the same classification. Consider the x-line oscillating gesture discussed earlier. Whether the gesture starts at the left side of the line or the right side (for example, x(0)=−1 or x(0)=+1), the resulting motions would still be identified by a human as the same gesture. Therefore, another type of representation seems desirable.
Since a human hand forms a gesture, we could imagine a representation in terms of the force exerted by the person's arm muscles. Alternatively, we might imagine representing the gesture as a function of the nerve impulses that travel from the brain to the arm's muscles. However, quite clearly, most of the countless types of such “internal” representations are presently impossible to quantify in any useful manner. Four hundred years ago, Newton developed a parsimonious representation of physical motions based on their dynamic properties,
{dot over (x)}(t)=f(x)
A dynamic system is a mathematical model describing the evolution of all possible states in some state space as a function of time [Hirsch 74] [Arnold 78]. The set of all possible states is a state space. Given an initial state, the set of all subsequent states as it evolves over time is a “trajectory” or “motion”. For any initial condition, the future evolution of the states in a trajectory remains within that trajectory (the trajectory is an invariant set). Thus, all that is required to describe a particular spatial motion is the differential equation representation and its initial conditions. We use a deterministic representation, as opposed to a stochastic one, because we believe these oscillatory motions are best represented by sine waves or a sum of exponentials as opposed to characteristics based on statistical properties. As with the geometric representation, there are an infinite number of gesture classifications of the form {dot over (x)}(t)=f(x). However, as before, we can choose a vector of tunable parameters to make the number of gesture classifications finite. Such representation has the form:
{dot over (x)}(t)=f(x,θ)
where θ represents the tunable parameters. Fixing the value of θ in a given representation yields a unique set of motions, with different initial conditions, described by {dot over (x)}(t)=f(x,θ). Motivated by the way humans interpret gestures, we associate an entire set of motions with one specific gesture. Thus, choosing different values of θ in a given representation results in a “family” of trajectories sets—a “gesture family.” For example, consider a oscillatory line gesture, the motion of which is constrained to the x-axis. This gesture can be represented in the following two dimensional state space:
{dot over (x)}1=x2
{dot over (x)}2=θ1x1
where x1 represents the position of the gesture, x2 is its velocity, and θ1 is a specified negative parameter. For any constant θ>0, all trajectories satisfy −θ1x12+x22=const as can be seen by direct differentiation.
We conceive of a specific gesture as a family of sets of trajectories. Referring to the gesture in
We will now represent a given family of gestures (family of sets of trajectories) by a mathematical model which contains a finite number of tunable parameters. A mathematical model described by differential equations, as above, allows the development of a computational scheme that will determine which parameters, the values of θi's, correspond to a specific gesture. The set of all valid parameters is the parameter space. The parameter space defines the family of gestures which can be represented by the model. In order to categorize a finite number of gestures in this family and to permit further variability in the exact motions associated with a particular gesture within this family, we partition the parameter space into a finite number of cells—the “lexicon”—and associate all the parameter values in the same cell with one gesture.
We will “invent” certain differential equations, composed of state variables and parameters, which intuition suggests may represent human gestures. Such differential equation models can be divided into two types: non-linear-in-parameters (NLIP) and linear-in-parameters (LIP). The two models can be further subdivided into linear-in-state (LIS) and non-linear-in-state (NLIS). It is advantageous to use a NLIP (with NLIS) model because it covers, by definition, a much broader range of systems than an LIP model. However, for reasons to be discussed below, we find it expedient to use a LIP model for our gesture representation. We have chosen to represent planar oscillatory gestures as a second order system with the intuition that a model based on the acceleration behavior (physical dynamics) of a system is sufficient to characterize the oscillatory gestures in which we are interested. This system's states are position and velocity. However, the vision system we use to sense gestures yields only position information. Since velocity is not directly measured, then either the parameter identification method could be combined with a technique for observing the velocity, or the velocity could be determined through position differences. In this section we show techniques for determining gesture parameters both when the velocity state is observed, and when it is obtained through position differences. By examining the utility of each technique, we develop an appropriate form of the gesture model and parameter identification method.
A difficulty with using human created gestures is that the underlying true physical model is unknown. Also, because people cannot precisely recreate even a simple circular gesture, multiple sets of parameters could represent the same gesture. Simulations are used both to determine a viable gesture model and to determine if it is possible to discover appropriate parameters for each gesture despite variations in motion.
We had chosen to represent motion gestures using dynamic systems. Our next task was to determined a model and a method for computing the model's parameters such that the model's parameters will best match an observed gesture motion.
Our choice of a model and parameter determination scheme was based on an exploration of the following issues:
We chose our gesture model and parameter determination scheme as follows. First, we decided to abandon off-line batch techniques in favor of on-line ones for reasons already discussed above. The on-line method needs to be chosen carefully, because there are relatively few cases where it can be guaranteed that the estimated parameters will be equivalent to those resulting from off-line techniques applied to the entire data set as a whole.
Next, in an attempt to use only position data, we examined a Series-Parallel Observer, which provides an estimate of the other unknown state for purely LIS and LIP systems. We abandoned this observer because it cannot adequately estimate parameters of non-perfect human gestures. Specifically, we could not extend the method to NLIS systems.
An on-line gradient descent method was examined, but for presently available methods applicable to NLIP systems, there is no guarantee that the parameters will converge towards their optimal values. Also, the parameters computed via this method are dependent on the order the data is presented.
A Linear Least Squares method (LLS) was examined next, which makes use of all the data independent of ordering. The resulting recursive LLS technique work for NLIP models, and, therefore, allow us to examine more flexible and useful gesture models. [See Cohen 96 for a detailed discussion of the reasons for the above decisions.]
The Recursive Linear Least Squares incrementally incorporates new data for determining the parameters which will best fit a set of data points to a given linear model. The recursive LLS method uses a tuning rule for updating the parameter vector θ without inverting a matrix, creating a more computationally efficient LLS algorithm. A tuning rule is required, because each block of data will result in a different set of parameters, as illustrated in
A method of incrementally updating the parameter θ is described below. For full details, see [Kumar 86]. The concept is illustrated in
The recursive (incremental) Linear Least Squares tuning method proceeds as follows. The tuning rule has the form:
θm+1=g(xm,{dot over (x)}m,θm)
Suppose we have the output data {dot over (x)} and state data x up to time in, and from this data we have already determined the best parameters θ for the set. From [Cohen 96] we know that at the next time step, with {dot over (x)}m+1 and xm+1:
This is an update law for the Rm+1 and θm+1 terms. We still have to find the inverse of Rm+1 at each time step. Fortunately, the matrix inversion lemma yields:
The above equation is a recursive formula for Rm+1−1 that is not based on taking the inverse of a matrix. The initial value of R0 is chosen as the identity matrix. See [Cohen 96] for a more formal discussion. If more importance is attached to recent data than to data received in the remote past, then we can choose θm to minimize:
where λ is termed the forgetting factor and is chosen with 0<λ<1. This results in:
The above recursive equation is the identifier in our gesture recognition system. This identifier allows us to represent gestures using a NLIP model, with the parameters identified using an on-line computationally efficient data order independent technique. We now determine the specific model used to represent oscillatory motion gestures.
Given that we modeled gestures using an LIP/NLIS representation, the following algorithm was used to determine the appropriate model (see
For the first step, we created phase-plane plots of the gestures to be modeled, as illustrated in the last plot in
{dot over (x)}1=x2
{dot over (x)}2=θ1x1+θ2
Intuition was used to “guess” appropriate models that would best match the phase plane motions. Because we believed an acceleration model will sufficiently characterize the gestures in which we are interested, the {dot over (x)}2 equation is the one modified with additional terms and parameters. For each model, the specific parameters for each gesture in the lexicon were computed using the LLS method (as discussed earlier).
The models were tested in simulation by measuring how well each tuned parameter model can predict the future states of its associated gesture (i.e. computing a total residual error). The model which best discriminates between gestures was chosen.
If none of the models can clearly discriminate between different gestures in a lexicon, then new models are required. The heuristic we used was to add or delete specific terms, and determine if there was a significant change (good or bad) in the model's ability to discriminate gestures. Adding two specific terms to the above equation, that is, using the new model
{dot over (x)}1=x2
{dot over (x)}2=θ1x1+θ2x2+θ3x2x12+θ4
results in a model that is better able to discriminate between gestures.
The results of this process of modeling oscillating circles and lines are detailed in the remaining parts of this section. This process will also be used in the Phase II effort to determine an appropriate model to classify certain non-linear gestures.
A variety of linear-in-parameter models for good circle and line gesture representations were tested. As before, each model represented only one dimension of motion, which was expanded to two or three for actually gesture recognition (i.e. an oscillating circle or line is formed when two or three of these decoupled models are present, one for each planar motion dimension). Again, x1 is the position state, and x2 is the velocity state. Five of these models are shown below. The determination of such models illustrates how a new (and more comprehensive model) could be determined when required for more complex dynamic motions.
To use the models described here on a digital computer, a fourth-order Runge-Kutta integration method was used [Press 88]. Simulations showed that a sampling rate of 10 Hz is sufficiently small to allow the use of this method.
The linear with offset component model is the most basic second order linear system. The offset component allows the model to represent gestures that are offset from the center of the image plane. It contains two parameters and is of the form:
{dot over (x)}1=x2
{dot over (x)}2=θ1x1+θ2
The Van der Pol equation is a slightly non-linear system, containing three parameters. The θ2 and θ3 parameters are attached to damping terms. This system is of the form:
{dot over (x)}1=x2
{dot over (x)}2=θ1x1+θ2x2+θ3x2x12
An offset component is added to the Van der Pol equation in this system. This system has four parameters and is of the form:
{dot over (x)}1=x2
{dot over (x)}2=θ1x1+θ2x2+θ3x2x12+θ4
A more non-linear system than the Van der Pol equations, the Higher Order Terms system contains additional spring-like components. This system has six parameters and is of the form:
{dot over (x)}1=x2
{dot over (x)}2=θ1x1+θ2x12+θ3x13+θ4x2+θ5x2x12+θ6
The Velocity Damping Terms system has additional damping terms. It contains eight parameters and is of the form:
{dot over (x)}1=x2
{dot over (x)}2=θ1x1+θ2x12+θ3x13+θ4x2+θ5x2x12+θ6x23+θ7x12x23+θ8
The use of simulations to determine the best gesture model for representing oscillating circles and lines is now detailed. We first details the residual measure calculation. Next the use of the residual measure to determine the best gesture model is described.
A predictor bin is composed of a model with parameters tuned to represent a specific gesture. The role of a bin is to determine a gesture's future position and velocity based on its current state. To measure the accuracy of the bin's prediction, we compared it to the next position and velocity of the gesture. The difference between the bin's prediction and the next gesture state is called the residual error. A bin predicting the future state of a gesture it represents will have a smaller residual error than a bin predicting the future state of a gesture it does not represent.
The computation for the residual error is based on equation:
{dot over (x)}k=FkTθ
Recall that f(x) is a two dimensional vector representing the gesture's position and velocity. Therefore {dot over (x)}k is the gesture's velocity and acceleration at sample k. We compute {dot over (x)}k from the gesture's current and previous position and velocity. The parameter vector {circumflex over (θ)} is used to seed the predictor bin. Then:
{circumflex over ({dot over (x)}k=fkT{circumflex over (θ)}
The residual error is then defined as the normalized difference between the actual value of {dot over (x)}k and the calculated value of {circumflex over ({dot over (x)}k:
The total residual error is the res_err summed for all data samples. The following section presents the residual calculation for each gesture with respect to each of the computed parameters.
We now detail how we determined which parameterization model for the predictor bin would best differentiate gestures. A data set of position and velocities of gestures is required to test each model. Using a vision system, data was recorded for a slow, medium, and fast circular gesture. The data is the x and y position and velocity measurements from the image plane of the vision system, although for these simulations only one of the dimensions is used. There is a small transition time when a human begins a gesture. This transient is usually less than a second long, but the residual error measurement has no meaning during this time. Therefore, gestures that last at least five seconds are used. The data recorded from a human gesture is termed “real gesture data.”
The total residual error was calculated by subjecting each predictor bin to each gesture type. A measure of a model's usefulness is determined by examining the ratio of the lowest residual error to the next lowest residual error in each column. The worst “residual error ratio” is the smallest ratio from all the columns because it is easier to classify a gesture when the ratio is large.
The residual error results of the Linear with Offset Component are shown in Table 1. The residual errors for the slow and medium gestures, with respect to their associated bins, are an order of magnitude lower than the other errors in their columns. The residual error of the fast gesture, with respect to the fast gesture bin, is one-forth the size of the closest residual error in its column (the medium gesture bin). Therefore, the Linear with Offset Component system is a good candidate for a gesture model.
As seen in Table 2, the Van der Pol model is only a fair candidate for gesture discrimination. The residual error of the medium gesture with respect to its gesture bin is only two-fifths smaller than the residual error with respect to the slow gesture bin. Also, the residual errors in the slow gesture column are not an order of magnitude apart.
The Van der Pol with Offset Component model is better at discriminating gestures than the model without the offset term (see Table 3). The residual errors in the medium gesture's column are now an order of magnitude apart. Although the residual errors in the fast gesture's column are not, the discrimination is still slightly better than in the Linear with Offset Component model.
Table 4 shows the residual errors associated with the Higher Order model. This model is an improvement over the Van der Pol with Offset Component model, as the residual errors in the fast gesture's column are now almost an order of magnitude apart.
Finally, Table 5 lists the residuals errors for the Velocity Damping model. This is the best model for discriminating between gestures, as the residual errors for each gesture with respect to their tuned bins are all at least an order of magnitude below the other residual errors in their columns.
A comparison of the worst “residual error ratio” of each model we considered is summarized in
Combining one-dimensional motions to form higher dimensional gestures.
We have shown how predictors can be used to recognize one-dimensional oscillatory motions. Recognition of higher dimensional motions is achieved by independently recognizing multiple, simultaneously created one-dimensional motions. For example, the combination of two oscillatory line motions performed in perpendicular axis can give rise to circular planar gestures, as shown in
Humans have the ability to create these planar motions. However, they can also make these motions in all three dimensions (for example, circles generated around different axis). To recognize these planar gestures, performed in three-dimensional space, a vision system must be able to track a gesture's position through all three physical dimensions. A binocular vision system has this capability, as does a monocular system with an attached laser range finder. Any of these such vision systems can be used with our gesture recognition system to identify three-dimensional gestures.
Development of a System to Recognize Static Gestures
Recognizing static hand gestures can be divided into localizing the hand from the rest of the image, describing the hand, and identifying that description. The module to recognize static hand gestures is to be both accurate and efficient. A time intensive process of evaluating hand gestures would prevent the system from updating and following motions which occur in real time. The system is intended to interact with people at a natural pace. Another important consideration is that the background may be cluttered with irrelevant objects. The algorithm should start at the hand and localize the hand from the surroundings.
Methodology
In order to meet these demands, the edges of the image are found with a Sobel operator. This is a very fast linear operation which finds approximations to the vertical and horizontal derivatives. In order to use only a single image, the greater of the horizontal and vertical component is kept as the value for each pixel. Besides being quick to calculate, an edge image avoids problems arising from attempting to define a region by locating consistent intensity values or even consistent changes in intensity. These values can vary dramatically in one hand and can be very hard to distinguish from the background as well.
In order to describe the hand, a box which tightly encloses the hand is first found. This allows a consistent description which is tolerant to changes in scale. To locate this box, we assume a point within the hand is given as a starting point. This is reasonable because the hand will be the principal moving object in the scene. Moving objects may be easily separated and the center of the largest moving area will be in the hand. From this starting point, a prospective box edge is drawn. If this box edge intersects an existing line, it must be expanded. Each side is tested in a spiral pattern of increasing radius from the initial center point. Once three sides have ceased expanding the last side is halted as well. Otherwise, the last side would often crawl up the length of the arm. The bounding box is shown in
Once the hand has been isolated with a bounding box, the hand is described (
The last task of the static gesture recognition is to identify the new description. A simple nearest neighbor metric is used to choose an identification. A file of recognized gestures is loaded in the initialization of the program. This file consists of a list of named gestures and their vector descriptions.
Considerations
The primary obstacle in static gesture recognition is locating and separating the hand from the surroundings. Using sophisticated models of the hand or human body to identify with an image are computationally expensive. If orientation and scale are not very constrained, this cannot be done in real time. Our system makes descriptions quickly and can compare them to predefined models quickly.
The limitations of the current system are a result of being dependent on the fast edge finding techniques. If lighting is highly directional, parts of the hand may be placed in shadow. This can cause odd, irregular lines to be found and defeat the normal description. If the background immediately surrounding the hand is cluttered with strongly contrasting areas, these unrelated lines may be grouped with the hand. This also causes unpredictable and unreliable descriptions. Such a background is very difficult to separate without making assumptions about the hand color or the size of the hand. An upper and lower bound are placed on the size of the hand in the image, but these permit a wide range of distances to the camera and are needed to assure that enough of the hand exists on image to make a reasonable description.
As long as the hand is within the size bounds (more than a speck of three pixels and less than the entire field of view) and the immediate surroundings are fairly uniform, any hand gesture may be quickly and reliably recognized.
Multiple camera views can be used to further refine the identification of static gestures. The best overall match from both views would be used to define and identify the static gestures. Furthermore, the system works not just for “hand” gestures, but for any static type of gestures, including foot, limb, and full body gestures.
The Overall Gesture Recognition System
In this section, based on the discussed functional and representational issues, we detail the specific components of our dynamic gesture recognition system from an architectural and implementational viewpoint. Our system is composed of five modules:
Gesture Creator
In the Gesture Creator module, a human or device creates a spatial motion to be recognized by the sensor module. Our gesture recognition system was designed to recognize consistent yet non-perfect motion gestures and non-moving static gestures. Therefore, a human as well as a device can create the gestures which can be recognizable by the system. Human gestures are more difficult to recognize due to the wide range of motion that humans recognize as the same gesture. We designed our gesture recognition system to recognize simple Lissagous gesture motions (repeating circles and lines), advanced motions such as “come here” and “go there”, and static hand symbols (such as “thumbs-up”).
Dynamic Gesture Lexicon
A gesture lexicon is a set of gestures used for communication or device control. We chose gestures for our lexicon based on the following:
The dynamic gestures used in this system consist of three one-dimensional oscillations, performed simultaneously in three dimensions (or two oscillations performed in two dimensions). A circle is such a motion, created by combining repeating motions in two dimensions that have the same magnitude and frequency of oscillation, but with the individual motions 90 degrees out of phase. A “diagonal” line is another such motion. To illustrate this, we define three distinct circular gestures in terms of their frequency rates: slow, medium, and fast. Humans create gestures that we define as slow large circles (slow), fast large circles (medium), and fast small circles (fast). More complex gestures can be generated and recognized, but these simple ones are used for illustrative purposes.
Main Three Gestures
Using the simpler Linear with Offset model (whose parameters are easier to understand than the more complex models), we represented a circle by two second order equations, one for each axis:
{dot over (x)}1=x2
{dot over (x)}2=θ1x1+θ2
and
{dot over (y)}1=y2
{dot over (y)}2=θ1y1+y2
Our gesture model has no “size” parameter. θ1 is a frequency measure, and θ2 is a drift component. The gestures were named “large”, “small”, “fast”, and “slow” due to the human motions used to determine the parameters (see
Expanded Lexicon—Geometric Constraints
A total of 24 gestures are possible from this example representation when the following are distinct gestures: clockwise and counter-clockwise circles, diagonal lines, one dimensional lines, and small and large circles and lines. Geometric constraints are required to expand the lexicon, because different gestures can result in the same parameters.
Given the various combinations of slow, fast, small, and large circles, the only one not used as a gesture is the slow small circle. Because, the slow small circle has the same oscillation frequency (medium) as the fast large circle, we need another geometric feature, the circle's size, to differentiate between these two gestures. As with the previous gestures, additional gestures can be created from these two gestures by varying the phase relationships.
Phase relationships are determined as follows. During the gesture, the x's and y's minimum and maximum image plane positions are computed. If the x and y motions are out of phase, as in a circle, then when x or y is minimum or maximum, the other axis's velocity is large. The clockwiseness of the motion is determined by looking at the sign of this velocity component. Similarly, if the x and y motion are in phase, then at these extremum points both velocities are small. A similar method is used when the gesture is performed in three dimensions.
Sensor Module
Unmodified Cohu Solid State CCD cameras are used as the sensor devices. No filters were used and the background was not modified. The Matrox Meteor capture card allows us to scale a captured image to any size without missing any frames. It will capture and transfer full-resolution, full-frame NTSC (640×480) or PAL (768×576) video input in real-time (30 Hz).
The color tracking system (CTS) uses the color of the hand and its motion to localize the hand in the scene. The hardware of the CTS system consists of a color camera, a frame grabber, and an IBM-PC compatible computer. The software consists of the image grabbing software and the tracking algorithm. Once the CTS is running, the graphical user interface displays the live image from the color camera on the computer monitor. The operator can then use the mouse to click on the hand in the image to select a target for tracking. The system will then keep track of the moving target in the scene in real-time.
The color tracking system is developed on a BSD 4.0 UNIX operating system. The hardware involved consists of a color camera, an image capture board and an IBM PC compatible. The software for the CTS is written in C and uses Motif for its graphical user interface (see
The present HTS system consists of a COHU 1322 color camera with a resolution of 494×768 pixels. The camera is connected to a Meteor image capturing board situated inside a Pentium-II 450 MHz IBM-PC compatible computer. The Meteor board is capable of capturing color video images at 30 frames per second. It is also able to capture these images at any resolution below the resolution of the camera.
The graphical user interface for the CTS displays a live color image from the camera on the computer screen. The user can then identify the target in the scene and click on it using the mouse. The CTS will then track the target in real-time. The flow chart of the tracking algorithm is shown in
We capture the image using functions from the Meteor driver. To provide real-time operation, we setup the board to signal the program using a system interrupt (SIGUSR2). Every time a new frame is ready, the Meteor alerts the program with an interrupt on this signal. The image capture function responds to the interrupt by transferring the current camera image to a buffer and processing it to find the target. The signal mechanism and its handling are what enable the system to operate in real-time. The graphical user interface of CTS displays the live camera image on the screen.
The user can start tracking by clicking the mouse on the target. This starts the tracking algorithm. The graphical user interface of the CTS is shown in
Once the user clicks on the target in the image, we compute the average color of a small region around this point in the image. This will be the color of the target region being tracked in the scene until it is reinitialized. Once tracking begins, we compute the position of the target region in the image using two methods. The first method tracks the target when there is sufficient motion of the target in the image. The second method will take over when there is no motion of the target in the scene.
Before choosing the methods for finding the target in the scene, the system checks for motion in a region near the current or estimated target position using a motion detecting function. This function computes the difference between the current image and the previous image, which is stored in memory. If motion has occurred there will be sufficient change in the intensities in the region. This will indicate motion. The motion detection function will trigger if a sufficient number of pixels change intensity by a certain threshold value.
If the motion detection function detects motion, the next step is to locate the target. This is done using the difference image and the target color. When an object moves between frames in a relatively stationary background, the color of the pixels changes between frames near the target (unless the target and the background are of the same color). We compute the color change between frames for pixels near the target location. The pixels whose color changes beyond a threshold make up the difference image. Note that the difference image will have areas, which are complementary. The pixels where the object used to be will complement those pixels where the object is at now. If we separate these pixels using the color of the target, we can compute the new location of the target. The set of pixels in the difference image, which has the color of the target in the new image, will correspond to the leading edge of the target in the new image. If we assume that the target approximates an ellipse of known dimensions, we can compute the position of the center of the target (ellipse) from this difference image (see
The color of a pixel in a color image is determined by the values of the Red, Green and Blue bytes corresponding to the pixel in the image buffer. This color value will form a point in the three-dimensional RGB color space (see
When the motion detection function fails to detect significant motion in the scene, we use a static target matching function to compute its location. The function searches a small area about the current location of the target to find the best fit in the image for the target. The search will find the location of the target with the highest matching value. We assume that the object is approximately elliptical. The elliptical target is hypothesized at each point in the search space and the matching metric is computed. This matching metric function uses a combination of edge and interior color matching algorithms to get a single matching number.
The image capture board is capable of providing us with a 480×640-pixel color image at 30 frames per second. Processing such a large image will slow down the program. Fortunately, the nature of the tracking task is such that, only a fraction of the image is of interest. This region called the window of interest lies around the estimated position of the target in the new image. We can compute the location of the target in the new image from the location of the target in the previous image and its velocity. This simple method is able to keep track of the target even when the target moves rapidly. We have found that the window of interest is typically 1/100th the area of the original image. This speeds up the computation of the new target location considerably. A computer with a higher processing speed could process the entire image without resorting to creating a region of interest.
Identification Module
The gesture recognition algorithms are located in the Identification Module. This module uses the position and velocity information provided by the sensor module to identify the gesture. The module is shown in
The Dynamic Gesture Prediction Module
The Dynamic Gesture Prediction module contains a bank of predictor bins (see
The idea for seeding each bin with different parameters was inspired by Narendra and Balakrishnan's work on improving the transient response of adaptive control system. In this work, they create a bank of indirect controllers which are tuned on line but whose identification models have different initial estimates of the plant parameters. When the plant is identified, the bin that best matches that identification supplies a required control strategy for the system [Narendra 94].
Each bin's model, which has parameters that tune it to a specific gesture, is used to predict the future position and velocity of the motion. This prediction is made by feeding the current state of the motion into the gesture model. This prediction is compared to the next position and velocity, and a residual error is computed. The bin, for each axis, with the least residual error is the best gesture match. If the best gesture match is not below a predefined threshold (which is a measure of how much variation from a specific gesture is allowed), then the result is ignored; no gesture is identified. Otherwise, geometric information is used to constrain the gesture further. A single gesture identification number, which represents the combination of the best x bin, the best y bin, and the geometric information, is outputted to the transformation module. This number (or NULL if no gesture is identified) is outputted immediately upon the initiation of the gesture and is continually updated.
The parameters used to initially seed each predictor bin were calculated by feeding the data of each axis from the three example basic gestures into the recursive linear least squares. The values for each bin are summarized in Table 6.
The Static Gesture Identification Module
The Static Gesture Identification module only searches for static gestures when the hand motion is very slow (i.e. the norm of the x and y velocities is below a threshold amount). When this happens, the module continually identifies a static gesture or outputs that no gesture was found.
The static gestures may be easily expanded by writing new gesture descriptions to a configuration file. Each gesture is described by a name tag, width, height, x location, y location, base side, and three vectors (in this example, each consisting of 15 integers) describing the profile of the hand. Because profiles may be significantly different due to varying tilts of the hand, multiple descriptions of fundamentally the same gesture may be desired. The initial or last line may also be less reliable due to missing the contours of the hand edge image.
The following are example parameter files (see Table 7). In each the name string is followed by an arm side, width, height, x location and y location. The arm parameter is simply an integer corresponding to above, below, right, or left. The width and height are measured in pixels. The x and y location are 0 if the location is not important or +1 or −1 to restrict recognition of a gesture to one particular quadrant. The following three vectors are the extreme side (the end of the hand) then the top or left side followed by the bottom or right side. The determination of which side is being represented is determined by the arm side parameter. For example, if the base side is from below (as in the Halt gesture below) the first line is from above, then from the left, then from the right. Right and left refer to the overall image—not the facing of the imaged person.
Another method used for this part is to parameterize each part of the hand (palm, digits, and wrist) as a set of connected “blobs”, that is, three-dimensional shapes which are connected together geometrically. As before, a configuration file would be used to defile how these blobs are connected, with the vision system identifying the blobs which this module sticks together.
The Overall Determination Module
This “Which Gesture?” module takes input from both the Static and Dynamic Gesture modules. When the velocity is small, then a static gesture is observed. When the velocity is greater than a threshold value, then a dynamic gesture is observed. The gesture identified is continuously outputted, and can therefore change value over time (the output can even be that no gesture was identified). The gesture identified is sent to the transformation module.
Transformation Module
The transformation module take a gesture type as its input and determines what to do with it. In the case of this system, the gesture is converted to parameters which represent the static or dynamic gesture, which is sent to the system which uses this information to produce a response.
System Response
The gesture command can be used for a wide variety of purposes. These include:
It is important to note that these devices can be controlled using static gestures, dynamic gestures, or a combination of the two. Thus, there is more information available to these system from the gesture input device, thereby allowing for a greater ability for humans to command and control them.
The key features of our architecture are the prediction modules and the signal flow from gesture creation to system response. The other modules could be replaced with functionally equivalent systems without changing the structure of our architecture. For example, instead of a human, a robot could create the gesture. Alternatively, one could create the gesture using a stylus, with a graphics tablet replacing the vision system in sensor module S. The graphics tablet would output the x and y coordinates to the identification module I. Similarly, module R could be a robot, one as complex as a six degree of freedom robot arm or as simple as a stepper motor based camera platform. The former mechanism requires a more complex transformation scheme in module T, while the latter system needs only a simple high level command generator.
As discussed earlier, the static and dynamic identification modules contains the majority of the required processing. Compared to most of the systems developed for gesture recognition (for example, see [Darrell 93] and [Murakami 91], for more details, see [Cohen 96]), this system requires relatively little processing time and memory to identify one gesture feature. This makes it possible to create a system with the ability to identify multiple features in parallel. A sophisticated module could then examine the parallel gesture features and infer some higher level motion or command.
The Behavior Recognition System
Just as the gesture recognition module is built on a bank of predictor bins, the behavior recognition system is composed of a bank of gesture recognition modules. Each module focuses on a specific point of the body (such as a foot or knee). As that point moves through space, a “gesture” is generated and identified. The combination of gestures from those points are what we define as a motion behavior, which can be categorized and identified. The system, illustrated in
Overall System Flow
The signal flow proceeds as follows. A user is tagged at various body locations automatically by the vision system. The data is acquired at 30 Hz and sent to a parser which splits off the data from each specific body location to its own gesture recognition module (GRM). There is one GRM for each tagged feature. Each GRM outputs which gesture it recognized (if any) which is again sent to an identification module. The identification module matches the gestures to their body location, defining a behavior. If this behavior matches one from a set of predefined behaviors, then this information is outputted.
The Parser
In the Parser module, the data, which is input as a stream of consecutive x,y,z, time coordinates from each tagged body location, is split up according to body location and sent to an appropriate GMI. This module needs to be changed whenever the input data is of a different format. Runtime variables define how many body parts are being tracked, and therefore the parser uses this information to determine the number of GMI bins and how to split up the data properly.
Gesture Recognition Modules (GRMs)
The time and coordinate data from each body feature is used as inputs to an appropriate GRM. Each GRM module is exactly as described in section II a, except that these modules handle three dimensional points instead of just two.
Behavior Identification Module
The Behavior Identification Module accepts as inputs gesture types and body locations from the GRMs. Various combinations of gestures at specific body locations are designated as behaviors, and if a match is found, then the program outputs that match.
Behavior Recognition Experiments and Data
We performed experiments to test the behavior recognition system. First, a variety of behaviors were performed and served as the baseline for identification. Then these behaviors were repeated and the data sent through the system.
The behaviors centered on repeated leg and waist motion. The three main types of behaviors performed were:
Jumping Jacks Start by standing straight, legs together, then move the feet out sideways and back together.
Walking in place: Move the feet in a slow walking motion.
Jogging in place: Move the feet up and down in an exaggerated quick motion.
Other behaviors used for experimentation, some of which included full three-dimensional motion:
Squats: Bend the knees to lower the body, then rise again.
Walking: Walk slowly towards the sensors.
Running: Run towards the sensors.
Skipping: Skip towards the sensors.
Hopscotch: Jump and land on one foot, then jump and land on both feet spread apart.
Rigorous experiments were performed using the first three behaviors. We used the Velocity Damping Terms gesture model to capture the richness of the presented motions. There was clear discrimination between the motions identified by the various identification bins, and, as shown below, behavior recognition was possible.
For the Jumping Jacks behavior, the eight theta's for each axis for each sensor are:
Sensor 1:
x-axis: 26.8199, 57.4879, 0, 0, 0, 1751.95, 0, 0,
y-axis: 0, 0, −257.759, 0, 0, 15.9921, 58.561, 0,
z-axis: 0, 0, 24.4205, 57.9981, 0, 0, 0, −1026.36,
Sensor 2:
x-axis: 17.8334, 58.4356, 0, 0, 0, 1691.1, 0, 0,
y-axis: 0, 0, −35.2513, 0, 0, 6.14527, 59.7759, 0,
z-axis: 0, 0, 28.6432, 57.4095, 0, 0, 0, 918.332,
Sensor 3:
x-axis: 15.0551, 58.7612, 0, 0, 0, 1186.79, 0, 0,
y-axis: 0, 0, −476.275, 0, 0, 7.1385, 59.4896, 0,
z-axis: 0, 0, 6.74113, 59.9307, 0, 0, 0, −544.907
Sensor 4:
x-axis: 10.8193, 58.9695, 0, 0, 0, −1210.42, 0, 0,
y-axis: 0, 0, 341.434, 0, 0, 9.92934, 59.1288, 0,
z-axis: 0, 0, 24.3505, 58.0358, 0, 0, 0, −142.869
Sensor 5:
x-axis: 16.7313, 58.3415, 0, 0, 0, 4060.06, 0, 0,
y-axis: 0, 0, 819.198, 0, 0, 15.3747, 58.6259, 0,
z-axis: 0, 0, 27.1073, 57.975, 0, 0, 0, 612.659.
The residual results are shown in Table 8. As anticipated, the bold diagonal cells have lower residuals than all the other cells in their respective rows for each gesture in the behavior, demonstrating that our system can discriminate this behavior type. Some of the residuals in a bin are very close in value to others sensors. This is because they are, in fact, the same gesture. For example, sensor 4 and sensor 5 are placed on the knee and foot of the same leg, which would make the same oscillatory motion during a behavior, so of course the residual values would be similar.
For the Walking behavior, the eight theta's for each axis for each sensor are:
Sensor 1:
x-axis: 24.304, 57.4632, 0, 0, 0, −351.126, 0, 0,
y-axis: 0, 0, −168.974. 0, 0, 23.2088, 57.6762, 0,
z-axis: 0, 0, 22.6047, 57.7623, 0, 0, 0, 1150.72,
Sensor 2:
x-axis: 19.8496, 57.8372, 0, 0, 0, −1017.67, 0, 0,
y-axis: 0, 0, 31.8642, 0, 0, 26.8075, 57.5024, 0,
z-axis: 0, 0, 25.6103, 57.9468, 0, 0.000123365, 0, −358.633,
Sensor 3:
x-axis: 10.2042, 59.182, 0, 0, 0, −617.508, 0, 0,
y-axis: 0, 0, 498.471, 0, 0, 30.6624, 56.95, 0,
z-axis: 0, 0, 22.534, 57.7824, 0, 0, 0, 598.156,
Sensor 4:
x-axis: 24.6263, 57.6805, 0, 0, 0, −73.7837, 0, 0,
y-axis: −0.000250812, 0, 125.269, 0, 0, 19.2459, 58.205, 0,
z-axis: 0, 0, 18.6133, 58.2271, 0, 0, 0, −195.928,
Sensor 5:
x-axis: 26.3399, 57.6672, 0, 0, 0, −522.552, 0, 0,
y-axis: −0.000136806, 0, 115.618. 0, 0, 18.8083, 58.326, 0,
z-axis: 0, 0, 19.0266, 58.1954, 0, 0, 0, 474.65,
The residual results are shown below in Table 9. Again, the bold diagonal cells have lower residuals than all the other cells in their respective rows for each gesture in the behavior, demonstrating that our system can discriminate this behavior type.
For the Running behavior, the eight theta's for each axis for each sensor are:
Sensor 1:
x-axis: 22.9112, 57.8263, 0, 0, 0, −1763.57, 0, 0,
y-axis: 0, 0, −489.467, 0, 0, 11.7958, 58.779, 0,
z-axis: 0, 0, −3.51229, 61.0138, 0, 0, 0, 713.328,
Sensor 2:
x-axis: −2.11517, 60.7157, 0, 0, 0, −40235.2, 0, 0,
y-axis: 0, 0, −4506, 0, 0, −79.0879, 70.5397, 0,
z-axis: 0, 0, −78.8084, 70.6087, 0, 0, 0, −375964,
Sensor 3:
x-axis: 24.5412, 57.6338, 0, 0, 0, −2805.13, 0, 0,
y-axis: 0, 0, −211.096, 0, 0, 23.1457, 57.6718, 0,
z-axis: 0, 0, 20.9598, 58.3911, 0, 0, 0, 773.77
Sensor 4:
x-axis: 20.1377, 58.218, 0, 0, 0, 4557.85, 0, 0,
y-axis: 0, 0, 607.713, 0, 0, 11.9292, 59.0339, 0,
z-axis: 0, 0, 16.2398, 58.6524, 0, 0, 0, −2667.72,
Sensor 5:
x-axis: 29.6411, 56.9948, 0, 0, 0, 1093.19, 0, 0,
y-axis: 0, 0, 954.695, 0, 0, 14.4107, 58.6439, 0,
z-axis: 0, 0, 20.9606, 58.0327, 0, 0, 0, 3497.27.
The residual results are shown below in Table 10. As before, the bold diagonal cells have lower residuals than all the other cells in their respective rows for each gesture in the behavior, demonstrating that our system can discriminate this behavior type. As for quadruped locomotion, Kelso [Kelso 95] hints that there is dynamic mechanical abstraction which can be studied to explain certain features of quadruped locomotion, specifically the limb frequencies of animals moving about the Serengeti plains. As he states: “When plotted against limb length or mass, the stepping frequency, from Thompson's gazelle to the black rhinoceros, falls on three straight lines, one for each locomotory mode.” Thus, it should be apparent to one of skill in the art that the invention is applicable to quadruped locomotion as well through appropriate extension.
Application Areas
Although one application of the invention is threat assessment, other uses are possible, including the generalized discrimination of suspicious or otherwise curious behaviors from normal activities. The system will not only perform surveillance activities, but make critical visual surveillance observations of locations for which there are not enough personnel to cover, or which are just simply too dangerous. Such surveillance tasks may include:
In the commercial area, this system provides the foundation for a large number of gesture recognition and behavior recognition applications. Following is a list of example commercial applications to which the technology is applicable:
ATM Control. Given our understanding of the gesture recognition, the gesture recognition system could be used for the control of Automatic Teller Machines (ATMs), as well as any other self-service machine or kiosk.
Crime Prevention. As discussed earlier, combinations of body gestures can be viewed as motion gaits, which can in turn be possibly identified as certain physical activities. Therefore, the same camera that is detecting gestures for the ATM (and any other self-service machine or kiosk) control can be used to identify the physical activity in the area. Such activities could include possible robberies or assaults. Having such crime detection features in place would greatly increase the safety of using an ATM.
General User Interfaces. This gesture and behavior recognition system can be incorporated into general user interfaces, to augment other input devices such as the mouse, keyboard, and voice commands. A system which could recognize behaviors would be useful in training people to use such devices.
As in the commercial arena, the gesture and behavior recognition system will have many uses in the military arena. The following is a list of military applications which we can expect to pursue in future projects.
Enhanced User Interfaces. As in the commercial application, the gesture and behavior recognition system can be incorporated into military computer systems that would benefit from an additional gestural user interface.
Surveillance. Since the dynamic gesture recognition can be used to detect physical activity at an ATM, it can also be used to detect activities in a combat environment. These activities not only include running, walking, and firing a weapon, but would also include the detection of group behaviors and formations.
This application is a continuation of U.S. patent application Ser. No. 11/412,252, filed Apr. 25, 2006, which is a continuation of U.S. patent application Ser. No. 09/540,461, filed Mar. 31, 2000, now U.S. Pat. No. 7,036,094, which claims priority of U.S. provisional application Ser. No. 60/127,510 filed Apr. 2, 1999. U.S. patent application Ser. No. 09/540,461 is a continuation-in-part of U.S. patent application Ser. No. 09/371,460, filed Aug. 10, 1999, now U.S. Pat. No. 6,681,031, which claims priority from U.S. Provisional Patent Application Ser. No. 60/096,126, filed Aug. 10, 1998. The entire content of each application and patent are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
4746770 | McAvinney | May 1988 | A |
5025314 | Tang et al. | Jun 1991 | A |
5047952 | Kramer et al. | Sep 1991 | A |
5252951 | Tannenbaum et al. | Oct 1993 | A |
5270820 | Fellinger | Dec 1993 | A |
5423554 | Davis | Jun 1995 | A |
5454043 | Freeman | Sep 1995 | A |
5481454 | Inoue et al. | Jan 1996 | A |
5490069 | Gioutsos et al. | Feb 1996 | A |
5544050 | Abe et al. | Aug 1996 | A |
5563988 | Maes et al. | Oct 1996 | A |
5570301 | Barrus | Oct 1996 | A |
5574498 | Sakamoto et al. | Nov 1996 | A |
5581276 | Cipolla et al. | Dec 1996 | A |
5594469 | Freeman et al. | Jan 1997 | A |
5612719 | Beernink et al. | Mar 1997 | A |
5652849 | Conway et al. | Jul 1997 | A |
5659764 | Sakiyama et al. | Aug 1997 | A |
5668573 | Favot et al. | Sep 1997 | A |
5670987 | Doi et al. | Sep 1997 | A |
5684701 | Breed | Nov 1997 | A |
5696892 | Redmann et al. | Dec 1997 | A |
5699441 | Sagawa et al. | Dec 1997 | A |
5710833 | Moghaddam et al. | Jan 1998 | A |
5714698 | Tokioka et al. | Feb 1998 | A |
5732227 | Kuzunuki et al. | Mar 1998 | A |
5757360 | Nitta et al. | May 1998 | A |
5759044 | Redmond | Jun 1998 | A |
5767842 | Korth et al. | Jun 1998 | A |
5774591 | Black et al. | Jun 1998 | A |
5798758 | Harada et al. | Aug 1998 | A |
5801704 | Oohara et al. | Sep 1998 | A |
5802220 | Black et al. | Sep 1998 | A |
5813406 | Kramer et al. | Sep 1998 | A |
5828779 | Maggioni et al. | Oct 1998 | A |
5838365 | Sawasaki et al. | Nov 1998 | A |
5864808 | Ando et al. | Jan 1999 | A |
5864848 | Horvitz et al. | Jan 1999 | A |
5875257 | Marrin et al. | Feb 1999 | A |
5878151 | Tang et al. | Mar 1999 | A |
5880411 | Gillespie et al. | Mar 1999 | A |
5887069 | Sakou et al. | Mar 1999 | A |
5889236 | Gillespie et al. | Mar 1999 | A |
5889523 | Wilcox et al. | Mar 1999 | A |
5898434 | Small et al. | Apr 1999 | A |
5901246 | Hoffberg et al. | May 1999 | A |
5903229 | Kishi et al. | May 1999 | A |
5907328 | Brush II et al. | May 1999 | A |
5907852 | Yamada et al. | May 1999 | A |
5912980 | Hunke | Jun 1999 | A |
5917490 | Kuzunuki et al. | Jun 1999 | A |
5966129 | Matsukuma et al. | Oct 1999 | A |
5973732 | Guthrie | Oct 1999 | A |
5990865 | Gard | Nov 1999 | A |
6002428 | Matsumura et al. | Dec 1999 | A |
6002808 | Freeman | Dec 1999 | A |
6005549 | Forest | Dec 1999 | A |
6009210 | Kang | Dec 1999 | A |
6014167 | Suito et al. | Jan 2000 | A |
6031568 | Wakitani | Feb 2000 | A |
6035053 | Yoshioka et al. | Mar 2000 | A |
6035067 | Ponticos | Mar 2000 | A |
6061055 | Marks | May 2000 | A |
6066954 | Gershenfeld et al. | May 2000 | A |
6075557 | Holliman et al. | Jun 2000 | A |
6075895 | Qiao et al. | Jun 2000 | A |
6082764 | Seki et al. | Jul 2000 | A |
6088017 | Tremblay et al. | Jul 2000 | A |
6088019 | Rosenberg | Jul 2000 | A |
6104383 | Shipman | Aug 2000 | A |
6130964 | Marques et al. | Oct 2000 | A |
6137908 | Rhee | Oct 2000 | A |
6144366 | Numazaki et al. | Nov 2000 | A |
6154559 | Beardsley | Nov 2000 | A |
6163336 | Richards et al. | Dec 2000 | A |
6173066 | Peurach et al. | Jan 2001 | B1 |
6188777 | Darrell et al. | Feb 2001 | B1 |
6219032 | Rosenberg et al. | Apr 2001 | B1 |
6272231 | Maurer et al. | Aug 2001 | B1 |
6275213 | Tremblay et al. | Aug 2001 | B1 |
6298170 | Morita et al. | Oct 2001 | B1 |
6301370 | Steffens et al. | Oct 2001 | B1 |
6335977 | Kage et al. | Jan 2002 | B1 |
6353427 | Rosenberg | Mar 2002 | B1 |
6456728 | Doi et al. | Sep 2002 | B1 |
6529809 | Breed et al. | Mar 2003 | B1 |
6545670 | Pryor et al. | Apr 2003 | B1 |
6720949 | Pryor et al. | Apr 2004 | B1 |
6766036 | Pryor et al. | Jul 2004 | B1 |
6804396 | Higaki et al. | Oct 2004 | B2 |
7050606 | Paul et al. | May 2006 | B2 |
7460690 | Cohen et al. | Dec 2008 | B2 |
20010008561 | Paul et al. | Jul 2001 | A1 |
20010029416 | Breed et al. | Oct 2001 | A1 |
20020029103 | Breed et al. | Mar 2002 | A1 |
20020036617 | Pryor | Mar 2002 | A1 |
20020097223 | Rosenberg | Jul 2002 | A1 |
20030125855 | Breed et al. | Jul 2003 | A1 |
20030209893 | Breed et al. | Nov 2003 | A1 |
20050008193 | Toyama | Jan 2005 | A1 |
20060202953 | Pryor et al. | Sep 2006 | A1 |
Entry |
---|
Birchfield, S. “Elliptical Head Tracking Using lnstensity Gradients and Color Histogram,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Santa Barbara, California, pp. 232-237, Jun. 1998. |
Horn, B.K.P. “Robot Vision,” The MIT Press, Cambridge, Massachusetts, 1986. |
US. Air force Instruction, “Aircraft Cockpit and Formation Flight Signals,” May 1994. |
U.S. Army Field Manual No. 21-60, Washington, D.C., Sep. 30, 1987. |
Arnold, V., “Ordinary Differential Equations,” MIT Press, 1978. |
Cohen, C., “Dynamical System Representation, Generation and Recognition of Basic Oscillatory Motion Gestures and Applications for the Control of Actuated Mechanisms,” Ph.D. Dissertation, Univ. of Michigan, 1996. |
Frank, D., “HUD Expands Kiosk Program,” Federal Computer Week, Mar. 8, 1999. |
Hager, G., Chang, W., Morse, A.; “Robot Feedback Control Based on Stereo Vision: Towards Calibration-Free Hand-Eye Coordination,” IEEE Int. Conf. Robotics and Automation, San Diego, CA, May 1994. |
Hauptmann, A., “Speech and Gestures for Graphic Image Manipulation,” Computer Human Interaction 1989 Proc., pp. 241-245, May 1989. |
Hirsch, M. Smale, S., “Differential Equations, Dynamical Systems and Linear Algebra,” Academic Press, Orlando, FL, 1974. |
Kanade, T., “Computer Recognition of Human Faces,” Birkhauser Verlag, Basel and Stuttgart, 1977. |
Karon, P., “ Beating an Electronic Pathway to Government with Online Kiosks,” Los Angeles Times, Aug. 25, 1996. |
Link-Belt Construction Equipment Co., “Operating Safety: Crames & Excavators,” 1987. |
Turk, M. et al., “Eigenfaces for Recognition,” Journal of Cognitive Neuroscience, 3, 1, 71-86, 1991. |
Narendra, K. et al., “Improving Transient Response to Adaptive. Control Systems Using Multiple Models and Switching,” IEEE Trans. on Automatic Control, 39:1861-1866, Sep. 1994. |
Rizzi, A. et al., “Distributed Real-Time Control of a Spatial Robot Juggler,” IEEE Computer, 25(5), May 1992. |
Wolf, C. et al., “The use of hand-drawn gesetures for text editing,” Int. Journ. of Man-Machine Studies, 27: 91-102, 1987. |
Wolf, C. et al., A Taxonomic Approach to Understanding Direct Manipulation, Jour. of the Human Factors Society 31th Annual Meeting, pp. 576-580. |
Yuille, A., “Deformable Templates for Face Recognition,” Journ. of Cognitive Neuroscience, 3, 1, 59-70, 1991. |
Cohen, C. et al., Issues of Controlling Public Kiosks and other Self Service Machines using Gesture Recognition, Oct. 1998. |
Conway, L. et al., Video Mirroring and Iconic Gestures: Enhancing Basic Videophones to Provide Visual Coaching and Visual Control, no. date available. |
Cohen, C. et al., Dynamic System Representation of Basic and Non-Linear in Parameters Oscillatory Motion Gestures, Oct. 1997. |
Cohen, C. et al., Dynamic System Representation, Generation, and Recognition of Basic Oscillatory Motion Gestures, Oct. 1996. |
Cohen, C. et al., The Control of Self Service Machines Using Gesture Recognition, (Aug. 1999). |
Wren et al., Pfinder: Real-Time Tracking of the Human Body, IEEE Transactions on Pattern Analysis and Machine Intelligence, 19(7): 780-785, Jul. 1997. |
Birchfield, S. “Elliptical Head Tracking Using Instensity Gradients and Color Histogram,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Santa Barbara, CA, pp. 232-237, Jun. 1998. |
Number | Date | Country | |
---|---|---|---|
20140071037 A1 | Mar 2014 | US |
Number | Date | Country | |
---|---|---|---|
60127510 | Apr 1999 | US | |
60096126 | Aug 1998 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 11412252 | Apr 2006 | US |
Child | 13850896 | US | |
Parent | 09540461 | Mar 2000 | US |
Child | 11412252 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 09371460 | Aug 1999 | US |
Child | 09540461 | US |