1. Field of the Invention
The present invention relates generally to systems and methods for processing data relating to a user's athletic performance and more particularly to systems and methods for processing that data in conjunction with data relating to virtual terrain.
2. Description of the Related Art
Stationary exercise equipment (see DEFINITIONS section), such as stationary bicycles and treadmills, are conventional. It is conventional to use a virtual course (see DEFINITIONS section) in connection with stationary exercise equipment. More specifically, the user makes virtual progress through the virtual course corresponding to her athletic exertion on the stationary exercise equipment. In most, if not all, of these stationary exercise equipment systems, the exercise equipments settings (for example, resistance or incline) are determined based on terrain factors of the virtual course. The virtual course might be represented simply by a numerical “distance” readout, such as on a numerical display panel mounted to the stationary bicycle. Alternatively, the virtual course might be represented by a 2d or 3d graphic display.
Some examples of stationary exercise equipment systems are disclosed in: (i) U.S. Pat. No. 6,450,922 (“Henderson”); (ii) US patent application 2008/0015089 (“Hurwitz”); (iii) U.S. Pat. No. 7,166,064 (“Watterson”); (iv) 2003/0214530 (“Wang”); and (v) US patent application 2005/0075213 (“Arick”). One drawback of stationary exercise equipment based systems is that the exercise experience is stationary (see DEFINITION of “stationary exercise equipment”). Stationary exercise equipment tends to be much less engaging, competitive and/or enjoyable than non-stationary exercise (see DEFINITIONS section).
In some of these publications listed in the previous paragraph, such as Watterson, the system includes a communications network that allows multiple users to progress on the same virtual course, and to, in some sense compete with and/or pace each other. However, it is noted that this competition or pacing is limited because of the control that the virtual course exerts over the exercise equipment settings of each user's exercise equipment. On the one hand, external control of the settings of the exercise equipment, based on parameters of the virtual course, may make: (i) the experiences of separate user's more similar to each other; and/or (ii) any competition between different users more “fair.” On the other hand, this external control of the exercise equipment limits each user's freedom in operating her own exercise equipment, and may therefore make the exercise less pleasant compared to exercise equipment based exercise where the user is in total control of her exercise equipment.
In other conventional exercise systems, athletic exertion exerted by the user during non-stationary exercise is measured. Generally, athletic exertion exerted during non-stationary exercise is not measured by monitoring of exercise equipment, but rather by monitoring the user's body itself. For example, motion detectors, such as accelerometers, may be worn on the body or on clothing in order to directly detect body motion and indirectly detect a degree of athletic exertion. In this document: (i) non-stationary athletic exertion detection based, at least in part, on any monitoring of the user's body (for example, GPS location of a user's body) will be referred to as body-based exertion detection; and (ii) non-stationary athletic exertion detection based, at least in part, upon detection of motion (for example, position, velocity and/or acceleration) of a part of the user's body will be referred to as body-motion-based exertion detection.
Some publications that may disclose body-based exertion detection and/or body-motion-based exertion detection include: (i) U.S. Pat. No. 6,148,280 (“Kramer”); (ii) US patent application 2002/0024450 (“Townsend”); (iii) US patent application 2005/0037844 (“Shum”); (iv) US patent application 2006/0064037 (“Shalon”); and (v) US patent application 2006/0262120 (“Rosenberg”).
US patent application 2007/0287596 (“Case”) discloses an athletic performance monitoring system that includes body-based exertion detection and, more specifically, body-motion-based exertion detection. Case discloses the use of a virtual course, stating: “The illustrated display 600 shows two virtual athletes 606 and 610 on the same (or approximately the same) course or route. The second virtual athlete 610 may represent the same individual and correspond to data associated with two different performances on the same (or similarly located) route. As another alternative, the second virtual athlete 610 may represent a different individual from the first virtual athlete 606, and the two displayed events may represent the two athlete's head-to-head competition in a single event or their individual efforts on the same course staggered or otherwise at different times. A second indicator display region 612 also may be provided to display instantaneous values of various measured time, distance, physical, and/or physiological parameters associated with the second virtual athlete 610's athletic performance at locations along the route. Of course, the data for the two athletic performances may be obtained from any source(s) without departing from the invention. Additionally, users may independently animate the action from each virtual athlete 606 and 610 (e.g., by separately activating “play” buttons 604 and 614 in the illustrated example or taking other appropriate action), or the virtual athletes 606 and 610 may be activated simultaneously and virtually ‘race’ one another (e.g., by activating the ‘play all’ button 616 in the illustrated example, or taking other appropriate action).”
Case goes on to disclose: “[A]fter running (or otherwise covering) any route with a GPS monitor in an athletic performance monitoring system and method according to examples of this invention (or otherwise obtaining data relating to the route, e.g., from different data sources, different sensors, etc.), an athlete, trainer, or coach may use that data in conjunction with a computer (or other processing device) to automatically generate their own virtual version of the route and/or use this data to develop an adjusted race plan for the route. For example, after providing the necessary route data for the system and method (e.g., by running or otherwise moving over the course with the GPS monitor system and/or other portions of an athletic performance monitoring system according to the invention), the tracking points may be downloaded to a computer. Then, upon supplying a target completion time for the same route, the software would provide route adjusted split times for each mile and/or other segment distances, optionally taking into account the elevational changes over the route, prevailing wind directions, wind speed, and/or other factors that may influence the difficulty of one portion of the course versus another. Such systems and methods also could calculate target split times to reach specific course landmarks, like aid stations, the halfway point, the bridge, and so forth. Optionally, such systems and methods may be operated shortly before the race begins, to enable incorporation of the local conditions at the time of the race (e.g., wind speed, direction, etc.). Any desired algorithm may be used to determine appropriate split times from a target overall time without departing from the invention.” (Note: in the foregoing excerpts of Case, all reference numerals refer to the Case patent document.)
Case is focused on recording data of specific performances of individuals and then allowing people to compare performances via their data. However, Case discloses that users whose performances are being compared must be “on the same (or approximately the same) course or route.” For this reason, Case cannot be considered to receive a user's athletic exertion as an input and normalize this input for virtual terrain (see DEFINITIONS of “virtual terrain” and “normalize athletic exertion for virtual terrain”). More specifically, the Case system does not provide a virtual environment that allows for performances to be normalized regardless of the degree of similarity in the real world courses or real world routes of respective performances (whether at different times and/or by different individual users). The Case system requires the real world terrain that respective users actually travel through during their non-stationary exercise to be at least “approximately the same.” This is a serious drawback of the Case system.
In the above-quoted portion of Case, it is disclosed that projected splits are adjusted by inputting assumptions about effects of elevation, wind, etc. However, this disclosure of the use of terrain factors in the Case system still does not normalize an actual performance, based on real athletic exertion, for virtual terrain. Rather, the terrain factors are merely used to calculate a virtual pacing performance. Case does not disclose that actual athletic exertion (of a single user or multiple users) can be used as a basis for pacing, competition or any other reason, unless, as discussed above, the performances are made on the same or similar course. Again, this is because Case does not normalize athletic exertion for virtual terrain.
To put some of the foregoing ideas in other words, the Case system merely simulates the real world with a virtual course. This is good because by limiting the virtual course to something that reflects the real world, a user's athletic exertion does not need to be normalized in order to compare performances made at different times or in different (similar terrain) places. On the other hand, the Case system does not allow: (i) a performance to be evocatively mapped onto a place significantly different from where a user is; and/or (ii) meaningful competition between non-stationary users travelling over significantly different terrain.
Other systems of possible interest (which may or may not be prior art) include the following: (i) US patent application 2008/0086318 (“Gilley”); (ii) US patent application 2003/0227453 (“Beier”); (iii) U.S. Pat. No. 7,278,966 (“Hjelt”); (iv) U.S. Pat. No. 6,672,991 (“O'Malley”); (v) U.S. Pat. No. 6,497,638 (“Shea”); and (vi) MAP MY RIDE website (note: MAP MY RIDE may be subject to trademark rights in some jurisdictions) at www.mapmyride.com/tdf/ (published as of Jun. 23, 2009).
Description Of the Related Art Section Disclaimer: To the extent that specific publications are discussed above in this Description of the Related Art Section, these discussions should not be taken as an admission that the discussed publications (for example, published patents) are prior art for patent law purposes. For example, some or all of the discussed publications may not be sufficiently early in time, may not reflect subject matter developed early enough in time and/or may not be sufficiently enabling so as to amount to prior art for patent law purposes. To the extent that specific publications are discussed above in this Description of the Related Art Section, they are all hereby incorporated by reference into this document in their respective entirety(ies).
According to at least some embodiments of the present invention: (i) a user's non-stationary exertion is detected; (ii) a virtual terrain is provided; and (iii) the detected non-stationary exertion is normalized for the virtual terrain. This is different than any stationary exercise machines that may normalize exertion for a virtual terrain because non-stationary exertion is used. This allows the user to exercise in a non-stationary way, such as by running cross-country. Non-stationary exercise is an advantage of at least some embodiments of the present invention because users often prefer non-stationary exercise to stationary exercise. At least some embodiments of the present invention make it possible to use mapping onto virtual terrains possible even in the non-stationary context. It is further noted that the present invention uses a virtual terrain (see DEFINITIONS section), rather than a virtual course (see DEFINITIONS section) without virtualized terrain factors (see DEFINITIONS section). By virtualizing terrain factors into a virtual terrain, at least some embodiments of the present invention provide a more realistic, realistically competitive and/or pleasing exercise experience for the user(s). By using both non-stationary exertion detection and a virtual terrain in combination exercise can be made more fun, competitive, pleasant and/or less unpleasant.
According to a first aspect of the present invention an exercise system for use in exercise by at least a first user includes non-stationary exercise equipment, a virtual terrain module and a normalization module. The non-stationary exercise equipment is structured and/or programmed to detect athletic exertion of the first user and output athletic exertion data corresponding to the detected athletic exertion. The virtual terrain module is structured and/or programmed to store virtual terrain data representing virtual terrain. The normalization module is structured and/or programmed to receive virtual terrain data from said virtual terrain module, to receive the athletic exertion data from the non-stationary exercise equipment, to determine at least one athletic exertion function based on the athletic exertion data, and to normalize the athletic exertion function for the virtual terrain data to determine virtual progress data corresponding to virtual progress through the virtual terrain.
According to a further aspect of the present invention, an exercise system for use in exercise by at least a first user and a second user, the system includes a first non-stationary exercise equipment, a second non-stationary exercise equipment, a virtual terrain module, and a normalization module. The first non-stationary exercise equipment is structured and/or programmed to detect athletic exertion of the first user and output first athletic exertion data corresponding to the detected athletic exertion. The second non-stationary exercise equipment is structured and/or programmed to detect athletic exertion of the second user and output second athletic exertion data corresponding to the detected athletic exertion. The virtual terrain module is structured and/or programmed to store virtual terrain data representing virtual terrain. The normalization module is structured and/or programmed to receive virtual terrain data from said virtual terrain module, to receive the first athletic exertion data from the first non-stationary exercise equipment, to determine a first athletic exertion function based on the first athletic exertion data, to normalize the first athletic exertion function for the virtual terrain data to determine first virtual progress data corresponding to the first user's virtual progress through the virtual terrain, to receive the second athletic exertion data from the second non-stationary exercise equipment, to determine a second athletic exertion function based on the second athletic exertion data, to normalize the second athletic exertion function for the virtual terrain data to determine second virtual progress data corresponding to the second user's virtual progress through the virtual terrain.
According to a further aspect of the present invention, an exercise system for use in exercise by at least a first user and a second user includes first exercise equipment, second exercise equipment, a virtual terrain module and a normalization module. The first exercise equipment is structured and/or programmed to detect athletic exertion of the first user and output first athletic exertion data corresponding to the detected athletic exertion. The second exercise equipment is structured and/or programmed to detect athletic exertion of the second user and output second athletic exertion data corresponding to the detected athletic exertion. The virtual terrain module is structured and/or programmed to store virtual terrain data representing virtual terrain. The normalization module is structured and/or programmed to receive virtual terrain data from said virtual terrain module, to receive the first athletic exertion data from the first exercise equipment, to determine a first athletic exertion function based on the first athletic exertion data, to normalize the first athletic exertion function for the virtual terrain data to determine first virtual progress data corresponding to the first user's virtual progress through the virtual terrain, to receive the second athletic exertion data from the second exercise equipment, to determine a second athletic exertion function based on the second athletic exertion data, to normalize the second athletic exertion function for the virtual terrain data to determine second virtual progress data corresponding to the second user's virtual progress through the virtual terrain.
The present invention will be more fully understood and appreciated by reading the following Detailed Description in conjunction with the accompanying drawings, in which:
Generally speaking, system 100 operates as follows: (i) a user puts on PDA 102, body mounted first sensor 120, mounting band 124 (with second sensor 122) and athletic shoe 125 (with third sensor 126); (ii) the user chooses a pre-existing virtual terrain from stored virtual terrains database 158 or defines a new virtual terrain through user input sub-system 115 and user interface sub-module 159; (iii) the user exercises in a non-stationary manner; (iv) the sensors 111, 120, 122, 126 of the sensor set detect non-stationary body motion (and/or other conditions relevant to non-stationary exercise); (v) the sensor data detected by the sensors is sent to normalization module 112 through sensor input module 113; (vi) selected virtual data is also sent from virtual terrain module 110 to the normalization module; (vii) the normalization module uses the sensor data and the virtual terrain data as inputs to determine the user's virtual progress through the virtual terrain; and (viii) the determined progress through the virtual terrain is communicated back to the user through user output sub-system 114.
In this embodiment, there is no network communication of data, but alternatively (or additionally) the sensor data, the virtual terrain data, user input data and/or user output data could be sent (by wireless and/or wired connection) through a communication network to one or more remote locations, such as a server computer, the user's home computer network, other user's at remote locations, etc.
Some (not all) possible variations on system 100 will now be mentioned and/or discussed. The PDA make take the form of a GOOGLE brand I-PHONE, and the modules and sub-modules may take the form of or include application(s) written for the I-PHONE (GOOGLE and/or I-PHONE may be subject to trademark rights in various jurisdictions throughout the world). The PDA could alternately be in the form of a home computer, a server computer, a dedicated computing/communication device (for example, a goggle mounted processing device, a relatively “dumb” terminal with a wireless transmission capability), etc. The various modules and sub-systems of PDA 102 do not need to be in a device carried around by the user during exercise. For example, the sensors could communicate to a server computer/home computer network, and the various modules and sub-systems of PDA 102 could be located, remotely from the exercising user, in a distributed manner among the server computer and the user's home computer. The GPS module may be omitted. For example, global positioning data may not be particularly helpful in some non-stationary exercise environments (like a swimming pool). Even if the GPS module is present, it does not need to be built into the PDA and could be a separate component.
User output sub-system 114 preferably includes audio output and a visual display. However, in embodiments where the user output sub-system is not included in equipment to be carried by the user during exercise: (i) the user will not be able to be informed of her ongoing progress through the virtual terrain on an ongoing basis as she is performing the exercise, but (ii) see may still watch her progress through the virtual terrain later, on a time delayed basis, when she gets to the location of the user output sub-system (for example, the monitor and speakers of her home computer). User input sub-system 114 preferably includes a keyboard and pointing device (for example, a touchscreen keypad and trackpad). However, in embodiments where the user input sub-system is not included in equipment to be carried by the user during exercise, the user may still provide user input (such as definition of a virtual terrain) before embarking on the non-stationary exercise.
Many variations on the sensors are possible. For example, in some embodiments the only sensor to sense body movement may be the GPS module. The GPS data may include both latitude/longitude type location, along with elevation, which is actually quite a significant amount of information with respect to some common exercises, like cross-country running. As another example, in some embodiments, the only sensor may be a single shoe mounted accelerometer. Although many preferred sensors are mounted to a user's body, sensors may also be mounted to non-stationary exercise equipment. One example would be a load cel operatively connected to a bicycle. Another example would be an air speed gauge mounted on a bicycle. Another example would be an altimeter mounted on a bicycle. Another example would be a current speed detector mounted on a rowboat. Generally speaking, more sensors and more sensor data can provide a better, more accurate estimation of a user's exertion due to the non-stationary exercise. For example, a sensor set that returns only the linear distance travelled by the user will generally lead to a less accurate estimate of exertion than a sensor set that accounts for linear distance, changes in elevation, changes in ground condition, wind speed, wind direction, oxygen content of the air, air pressure and so on. The cost of providing various additional kinds of sensors should be weighed against their utility in helping to make a more realistic real/virtual exercise experience. The more realistic the approximation of exertion, the more realistic competition can be, which is especially advantageous in multi-user systems, as discussed in greater detail below.
Referring now to
Course types sub-module 150 provides for different types of courses, such as running/walking courses, cross-country skiing courses, rowboat courses, bicycle courses, distance swimming courses and so on. Depending on the embodiment, different types of courses may look for different types of detected exertion as input data from sensors. For example, a swimming virtual terrain may look for exertion in terms of accelerations of arms and legs as inputs, while a cross country skiing course may supplement arm and leg acceleration data with sensor input data concerning snow condition and relative elevation. As a further example, some embodiments may treat all physical exertion as equivalent for purposes of tracking progress for the course. In this example, a jogger (in the real world) may be charted as if he were rowing a virtual rowboat, with the exertions of the jogger's real world legs being appropriately normalized to approximate arm exertions of a virtual rower. In some embodiments, course type module 150 will be restricted to a single course type.
After a course type is selected, appropriate terrain factors are added in, in order to make the virtual course into a virtual terrain (see DEFINITIONS section). In system 100, these terrain factors come from terrain sub-modules 151, 152, 153, 154, 155, 156, 157. These terrain factors may be set by user(s), pre-programmed or they may even be sensed by sensing equipment. For example, GPS data, temperature data and wind speed data collected during an actual bicycle trip taken by a real world user may be used to define a virtual terrain for the use of that particular user or for other users in other places. As a further example, GPS data, temperature data and wind speed data collected during an actual bicycle trip taken by a real world user, supplemented with terrain factor additions or modifications made by user(s) based on user input data may be used to define a virtual terrain.
Course geometry sub-module 151 can provide a terrain factor by giving a virtual course elevation values. This type of terrain factor is especially useful for walking courses, running courses and bicycle courses. Other types of geometry-related terrain factors may be provided. For example, a virtual swimming pool course may specify that opposite ends of the poll are some given virtual distance apart. As a further example, turns in an ice skating course may be considered as important terrain factors, even though elevation of the ice skating course is constant. Some virtual geometrical terrain factors may not even exist in the real world. For example, a virtual course could include a hyperspace portal that the user can choose to virtually enter in order to instantly put herself further ahead or further behind on a virtual terrain.
Ground condition sub-module 152 provides a virtual terrain factor in terms of ground condition, such as wet, icy, slippery, spongy, mushy, hard, flat and the like. For example, a user may record a realistic pacing performance for later use on a dry day even if the day of the pacing performance is a wet day. Preferably, any ground condition that significantly affects performance or the relationship between exertion and progress over the course is included in the ground condition sub-module. The ground conditions may be constant over the entire virtual terrain, or they may be defined to be different on different portions of the virtual terrain.
Water condition sub-module 153 provides virtual terrain factors relating to water condition for water-based kinds of courses. Water condition terrain factors may include current, turbulence or salinity. Preferably, any water condition that significantly affects performance or the relationship between exertion and progress over the course is included in the water condition sub-module. The water conditions may be constant over the entire virtual terrain, or they may be defined to be different on different portions of the virtual terrain.
Weather condition sub-module 154 provides virtual terrain factors relating to water condition for various kinds of outdoor courses. Weather condition terrain factors may include rain, wind, snow, blowing dust and hail. Preferably, any weather condition that significantly affects performance or the relationship between exertion and progress over the course is included in the weather condition sub-module. The weather conditions may be constant over the entire virtual terrain, or they may be defined to be different on different portions of the virtual terrain.
Light condition sub-module 155 provides virtual terrain factors relating to light conditions for various kinds of outdoor courses. Light condition terrain factors may include visibility and incoming solar radiation. Preferably, any light condition that significantly affects performance or the relationship between exertion and progress over the course is included in the light condition sub-module. The light conditions may be constant over the entire virtual terrain, or they may be defined to be different on different portions of the virtual terrain.
Temperature condition sub-module 156 provides virtual terrain factors relating to temperature condition for various kinds of courses. Temperature condition terrain factors may include absolute air temperature, absolute water temperature, temperature relative to the dew point and wind chill factor. Preferably, any weather condition that significantly affects performance or the relationship between exertion and progress over the course is includes in the ground condition sub-module. The temperature conditions may be constant over the entire virtual terrain, or they may be defined to be different on different portions of the virtual terrain.
Atmosphere condition sub-module 157 provides virtual terrain factors relating to temperature condition for various kinds of courses. Temperature condition terrain factors may include absolute air temperature, absolute water temperature, temperature relative to the dew point and wind chill factor. Preferably, any weather condition that significantly affects performance or the relationship between exertion and progress over the course is includes in the ground condition sub-module. The temperature conditions may be constant over the entire virtual terrain, or they may be defined to be different on different portions of the virtual terrain.
User interface sub-module 159 includes the logic and/or instructions that allow user(s) (local users or remote users connected through a communication network) to select, modify and/or define a new virtual terrain using the course type sub-module 150, the virtual terrain feature sub-modules 151 to 157 and/or the stored virtual terrains database 158. Alternatively, the user interface sub-module may be limited to only allow the user(s) to select a predetermined virtual terrain. Alternatively, the user interface sub-module may be limited to requiring the user to define any virtual terrain without the capability of providing predetermined virtual terrains. Alternatively, virtual terrains may be provided according to preprogrammed logic so that the user has no role in defining or selecting virtual terrains, which would render the user interface sub-module unnecessary.
Returning now to
The detection of the sensor inputs used to estimate exertion does not need to be done at the same time that exertion is estimated. For example, the sensor inputs, mapped against time, may be stored in a computer readable file for later use in estimating exertion, and normalizing that exertion for a virtual terrain, long after the user has completed her exercise.
The estimation of exertion does not need to be done at the same time that normalization is performed. For example, the estimated exertion, mapped against time, may be stored in a computer readable file for later use in performing normalization, long after the user has completed her exercise.
In some, not necessarily preferred embodiments, there may be no correspondence between the terrain factors of the virtual terrain and the sensor inputs used to estimate exertion. For example, consider an embodiment where the only sensor is a bicycle odometer, and exertion is estimated by making the simplified assumption that each meter travelled on the bicycle is equal to x calories of exertion. In this embodiment no data corresponding to real world virtual terrain factor(s) is used because mere distance is not a terrain factor (see DEFINITIONS section). Of course, despite the simplicity and inexpensiveness of the hardware required, this is not necessarily preferred because the estimate of athletic exertion is not likely to be an accurate estimate because real world terrain factors like elevation and wind speed are not considered in estimating exertion.
In other embodiments, there may be there may be strong correspondence between the terrain factors of the virtual terrain and the sensor inputs used to estimate exertion. For example, consider an embodiment where the sensor set includes: (i) a distance sensor; (ii) an elevation sensor; (iii) a wind speed/direction sensor; (iv) an oxygen sensor; (v) an atmospheric pressure sensor; and (vi) a user input module that allows the user to input ground conditions as she travels over a real world terrain. Consider further that the virtual terrain in this embodiment includes the following terrain factors: (i) an elevation; (ii) a wind speed/direction; (iii) oxygen richness of the air; (iv) atmospheric pressure; and (v) ground conditions. In this embodiment there is a strong correspondence between the factors used to estimate exertion and the virtual terrain factors of the virtual terrain. One good thing about this embodiment is that it will yield a realistic and fair normalization result in terms of progress of the user through the virtual terrain. This is especially helpful when competition is involved. One potential drawback of this embodiment is that many sensors must be acquired, connected, mounted and maintained.
Still other embodiments may have an intermediate degree of correspondence between the terrain-related sensor input data used to estimate exertion and the virtual terrain factors that are virtualized in the virtual terrain. For example, a cross-country skiing embodiment may use as sensor inputs for calculating estimated exertion: (i) left arm acceleration; (ii) right arm acceleration; (iii) right leg acceleration; (iv) left leg acceleration; and (v) GPS data including distance and elevation. This cross-country skiing embodiment may use as its virtual terrain factors: (i) elevation; and (ii) snow surface condition. In this embodiment, there is some correspondence between the sensor input (used to estimate exertion) and the virtual terrain factors (of the virtual terrain), but it is not a perfect correspondence.
Normalization module 112 outputs the normalized progress through the course to user output sub-system 114 for output (for example, audio output, video output) to user(s). The output to user(s) does not need to be done at the same time the normalization is calculated. For example, the normalized progress, mapped against time and the virtual space of the virtual terrain, may be stored in a computer readable file for later presentation to user(s) upon demand. Exertions based on multiple users, or multiple exercise sessions of a single user, or both may be combined to form a single presentation output to user(s). For example, user A may run two real world marathons in different cities, and user B may cross country ski the same 26 mile course twice on consecutive weekends. Estimated exertions corresponding to each of these four exercise sessions may be normalized and combined to make a presentation where four virtual competitors, one corresponding to each exercise session, are competing in a wheelchair race in the Sea of Fecundity on the moon. A third user, user C, may watch, and supplement with her normalized exercise session, this presentation on her waterproof PDA while she swims the English channel, such that she becomes the fifth chair in the lunar wheelchair race. If user C grow tired of this virtual terrain when she is halfway to France, then she may take a moment to switch virtual terrains so that all of the four other virtual competitors are swimming a virtual version of the English Channel right there with her, effectively pacing her, bringing out her competitive spirit and spurring her to complete the difficult swim.
Another embodiment will now be discussed with some reference to
Sensor data is converted to the most basic components of time and distance in x, y and z directions. This is achieved by first establishing a point of reference. For example, in the case of accelerometer data with a runner, this is achieved by establishing the point where a runners foot first impacts the ground (whether this is the actual “footplant” or not, is not important as reliability and consistency of measurement across strides is of most importance). The current example further assumes that this data is collected by applications of relatively ubiquitous technologies, such as the Apple iPhone (numerous other means of collecting sensor data can be used). In order to collect reliable data in this case, the phone can be strapped to ones mid-section (either in the front, just below one's belly button or on their back on the flatter portion of one's back, or the region of the extensor flexor muscle). The mass of the phone, however, results in a delay of the phones actions to that of a user's because it is inevitably interfacing with soft tissue (whether it be directly or via a harness of some kind), thus (regardless of whatever harnessing is used) capable of being accelerated in any direction as an independent object. Because, however, the phone is moving relative to the user in this case, all extraneous motions of the device cancel out. That is, if it accelerated off of one's body, it will accelerate in the opposite direction in a similar manner, but both actions will happen relative to the user's movements. To account for this, each “footplant” is identified at specific point in time and space by the programming used (for example, C, C++, PHP, etc.). In this example, the reference point is identified by a value that moves below a given threshold. This value is created by a momentary and drastic decrease in g-force caused by the impact of one's body to a relatively rigid body such as the ground. Identification of the next footplant does not commence until a given period of data points has been collected (or a given period of time has elapsed). This is shown in the graph 300 of
Distances and time are then converted to velocity and incline. Graph 400 of
This is achieved by the following general equations (assuming data is given in terms of acceleration due to gravity, g, or roughly 9.80665 m/s2): (i) for time:
where x=length of array; (ii) for distance: (a) convert raw data into distance traveled per time between data points):
and, (b) calculate total distance between reference points:
where x=length of array; and (iii) for incline:
where x=length of array.
Work performed, thus velocities (meters/second), distances or times, are normalized given inclines of each user. This normalization is achieved across runners because one can assume body weight across users to be equal. Per a typical sporting event, one does not share the experience of another's body weight, but does share the relative and shared effects of gravity on their performance. This thus allows the effects of gravity based on incline to be used as a variable for normalizing speeds and distances across users.
If we assume that people perceive either work or power, then to “equate” experiences in a meaningful manner, we must equate either work or power generated across individuals or an individual's experience across time. That is, this preferred embodiment of a system according to the present invention: (i) allows meaningful shared experience across individuals in distributed spaces; and (ii) allows meaningful combining of experiences of an individual across time. One important aspect of an athletic experience that at least some embodiments focus on is the experience of work and/or power generation. The behavioral assumptions discussed above allow application of common physical relationships in order to achieve meaningful representations of experiences to multiple users across both space and time.
Some relevant equations include the following: (i) Force=mass(m)*acceleration(a); (ii) Work=m*a*distance(d); and (iii) Power=m*a*d/time(t). Assume m1, a1, d1 and t1 are the mass, acceleration, distance and time associated with User 1. Also assume that User 1 is the “lead user” discussed above. Further assume m2, a2, d2 and t2 are the mass, acceleration, distance and time associated with User 2. Also assume that User 2 chosen to participate in an event with User 1 as discussed above. Also, because of the experience of any user's mass (or weight) is not shared as discussed above, it can be assumed that m1=m2. In order for W1 and W2, or P1 and P2, to be equated in a personally meaningful manner to users, it is assumed that Work has two components, one which is associated with the amount of work associated with overcoming the effects of the earth's gravity (m*a*d) and a second which is associated with the amount of work generated by user other than that required to overcome the effects of gravity (m*a*d), thus associated with the more relative and perceived differences this method captures.
If, in addition to all assumptions noted above, it is further assumed that W1=(F1+Fg)*d1 and that W2=(F2+Fg)*d2, then P1=W1/t1 and P2=W2/t2, or:
Where θ1 equals the incline of User 1, θ2 equals the incline of User 2 and d′2 equals the adjusted (virtual) distance for User 2.
Assumptions about mass discussed above are applied and it is assumed that t1=t2:
Alternatively, one might also solve for θ′ or t′ in a similar manner. That is, by equating Work or Power of two individuals and solving for an “unknown” variable given the known and/or measured values of all other variables. Also, note, that while we assume mass to be equal, it is possible for users to enter their weight and treat it as a known variable in the above equations.
All normalization can be performed based on relative work of each user to every other user. This can be achieved by having every event include a single user and terrain to which all other users and terrains will be compared. Work can be compared across users and a virtual distance, velocity and/or time is calculated. This number can then be used to populate every user's data file with both real performance data and terrain and their avatar's performance and terrain, or virtual performance and terrain. This then allows the system to properly position every user's avatar relative to every other avatar in a given event within a virtual environment. Note the connection back to one's actual experience allows the personal experience (that is., the experience any user can achieve through exercise absent any effects of or needs for technology) to be preserved. Leaving one to perform in both the real and virtual environments by virtue of single sporting effort. This can be represented through various means such as the one shown in screenshot 500 of
The fundamental data of every user is then stored in more elemental spatial and temporal forms, thus allowing the data to be used/recombined in all the synchronous and asynchronous manners discussed above. This also allows for the pure spatial element of terrain to be extracted from any data set and recombined with temporal and spatial data collected previously or at a later time. This also allows for data sets specific to real and virtual experiences to be stored and reused at a later time.
In some embodiments of the present invention, a scenario involves two or more individuals who are either co-located or not co-located training or competing with each other in real or close-to real time. Sensors provide input/data from each individual to the virtual environment in which individual interact, train and compete. These sensors allow the relative positions of each individual to be normalized and portrayed in the virtual environment through the application of, for example, avatars. The “terrain” in the virtual environment can be (i) that being “recorded” by one of the individuals (e.g., agreed upon by the group), (ii) an environment generated by the computing system, or (iii) previously uploaded into the system, or (iv) uploaded to the system by another individual.
In some embodiments of the present invention, a scenario involves two or more individuals who are either co-located or not co-located training or competing with each other in real or close-to real time whereby some of the individuals are on exercise equipment or training simulators (e.g., a treadmill or on a rowing machine) and some are “in a natural environment” exercising in more typical locations (e.g., on a road or on the water in a rowing shell). Sensors provide input/data from each individual to the virtual environment in which individual interact, train or compete. These sensors allow the relative positions of each individual to be normalized and portrayed in the virtual environment through the application of, for example, avatars. The “terrain” in the virtual environment can be (i) that being “recorded” by one of the individuals (e.g., agreed upon by the group), (ii) an environment generated by the computing system, or (iii) previously uploaded into the system, or (iv) uploaded to the system by another individual.
In some embodiments of the present invention, a scenario involves two or more individuals who are either co-located or not co-located training or competing with each other in real or close-to real time whereby all of the individuals are on exercise equipment or training simulators (e.g., a treadmill or on a rowing machine). Sensors provide input/data from each individual to the virtual environment in which individuals interact, train or compete. These sensors allow the relative positions of each individual to be normalized and portrayed in the virtual environment through the application of, for example, avatars. The “terrain” in the virtual environment can be (i) that being “recorded” by one of the individuals (e.g., agreed upon by the group), (ii) an environment generated by the computing system, or (iii) previously uploaded into the system, or (iv) uploaded to the system by another individual. It is noted that the system of
In some embodiments of the present invention, a scenario involves one individual who performs and records their performance or training locally, then later uploads their performance to a computing environment, then downloads their performance to their mobile technologies, which then allows them to interact directly with their own previous performances or trainings. In this way, the individual is temporally separated from a previous experience, but allowed to interact in real time or close to real time with “themselves” (e.g., now represented by various avatars, one for each previous performance or training as well as one for their current performance). Sensors allow the relative positions of each performance to be normalized and portrayed in the virtual environment per scenarios described in previous figures. The “terrain” in the virtual environment can be (1) that being “recorded” at the current time, (2) an environment generated by the computing system, or (3) previously uploaded into the system by the individual themselves, or (4) uploaded to the system by another individual.
In some embodiments of the present invention, a scenario involves one individual who performs and records their performance or training locally, then later uploads their performance to a computing environment, then downloads their performance to their mobile technologies, which then allows them to interact directly with their own previous performances or trainings whereby the previous performance was “in a natural environment” and the current performance is on a piece of exercise equipment. In this way, the individual is temporally separated from the previous experience, but allowed to interact in real time or close to real time with “themselves” (e.g., now represented by various avatars, one for each previous performance or training as well as one representing the current experience). Sensors allow the relative positions of each performance to be normalized and portrayed in the virtual environment per scenarios described in previous figures. The “terrain” in the virtual environment can be (1) that being “recorded” at the current time, (2) an environment generated by the computing system, or (3) previously uploaded into the system by the individual themselves, or (4) uploaded to the system by another individual.
In some embodiments of the present invention, a scenario involves one individual who performs and records their performance or training locally, then latter uploads their performance to a computing environment, then downloads their performance to their mobile technologies, which then allows them to interact directly with their own previous performances or trainings whereby the previous performance was on a piece of exercise equipment and the current performance is on a piece of exercise equipment. In this way, the individual is temporally separated from a previous experience, but allowed to interact in real time or close to real time with themselves (e.g., now represented by various avatars, one for each previous performance or training as well as one representing the current experience). Sensors allow the relative positions of each performance to be normalized and portrayed in the virtual environment per scenarios described in previous figures. The “terrain” in the virtual environment can be (1) that being “recorded” at the current time, (2) an environment generated by the computing system, or (3) previously uploaded into the system by the individual themselves, or (4) uploaded to the system by another individual. It is noted that the system of
In some embodiments of the present invention, a scenario involves one individual who performs and records their performance or training locally, then latter uploads their performance to a computing environment, then downloads their performance to their mobile technologies, which then allows them to interact directly with their own previous performances or trainings whereby the previous performance was on a piece of exercise equipment and the current performance is “in a natural environment.” In this way, the individual is temporally separated from a previous experience, but allowed to interact in real time or close to real time with themselves (e.g., now represented by various avatars, one for each previous performance or training as well as one as well as one representing the current experience). Sensors allow the relative positions of each performance to be normalized and portrayed in the virtual environment per scenarios described in previous figures. The “terrain” in the virtual environment can be (1) that being “recorded” at the current time, (2) an environment generated by the computing system, or (3) previously uploaded into the system by the individual themselves, or (4) uploaded to the system by another individual.
In some embodiments of the present invention, a scenario involves one individual who performs and records their performance or training locally, then latter uploads their performance to a computing environment designed to allow other individuals to download their experience (i.e., performance or “terrain” or both), thus allowing them to interact directly with a recorded performance or in specific terrain albeit normalized in the virtual environment. The “terrain” in the virtual environment can be (1) that being “recorded” at the current time, (2) an environment generated by the computing system, or (3) previously uploaded into the system by the individual themselves, or (4) uploaded to the system by another individual.
In some embodiments of the present invention, a scenario involves one individual who performs and records their performance or training locally, then latter uploads the terrain trained or performed on to a computing environment, then downloads the terrain to their mobile technologies, which then allows them to train on previous terrain. In this way, the individual is not co-located on the previous terrain, but allowed to interact in real time or close to real time as if they were via the normalization performed by the system. This “terrain” could potentially include all factors that might impact performance (e.g., percent grade/incline, wind, temperature, etc.). Sensors allow the relative positions of each performance to be normalized and portrayed in the virtual environment per scenarios described in previous figures.
In some embodiments of the present invention, a scenario involves one individual who performs and records their performance or training locally, then latter uploads the terrain trained or performed on to a computing environment in a computing space to which others have access, who then download the terrain to their mobile technologies, which then allows them to train on previously recorded terrain. Sensors allow the relative positions of each performance to be normalized and portrayed in the virtual environment per scenarios described in previous figures.
In some embodiments of the present invention, a scenario involves two or more individuals who are either co-located or not co-located training with each other in real or close-to real time whereby at least one member is a “coach.” Sensors provide input/data from each individual to the virtual environment in which individuals interact.
In some embodiments of the present invention, a scenario involves one individual or two or more individuals who are either co-located or not co-located training or competing with each other in real or close-to real time. Sensors provide input/data from each individual to the virtual environment in which individual interact, train and compete. These sensors allow the relative positions of each individual to be normalized and portrayed in the virtual environment through the application of, for example, avatars. The “terrain” in the virtual environment can be (1) that being “recorded” by one of the individuals (e.g., agreed upon by the group), (2) an environment generated by the computing system, or (3) previously uploaded into the system, or (4) uploaded to the system by another individual. In this way, as per previous scenarios, individual can create “virtual performances” based on actual performances of themselves, themselves and other individuals, or only other individuals within specific environments by cross referencing these components in an “on demand” manner.
Any and all published documents mentioned herein shall be considered to be incorporated by reference, in their respective entireties, herein to the fullest extent of the patent law. The following definitions are provided for claim construction purposes:
Present invention/invention: means at least some embodiments of the present invention; references to various feature(s) of the “present invention” throughout this document do not mean that all claimed embodiments or methods include the referenced feature(s).
Embodiment: a machine, manufacture, system, method, process and/or composition that may (not must) meet the embodiment of a present, past or future patent claim based on this patent document; for example, an “embodiment” might not be covered by any claims filed with this patent document, but described as an “embodiment” to show the scope of the invention and indicate that it might (or might not) covered in a later arising claim (for example, an amended claim, a continuation application claim, a divisional application claim, a reissue application claim, a re-examination proceeding claim, an interference count); also, an embodiment that is indeed covered by claims filed with this patent document might cease to be covered by claim amendments made during prosecution.
First, second, third, etc. (“ordinals”): Unless otherwise noted, ordinals only serve to distinguish or identify (e.g., various members of a group); the mere use of ordinals implies neither a consecutive numerical limit nor a serial limitation.
Electrically Connected: means either directly electrically connected, or indirectly electrically connected, such that intervening elements are present; in an indirect electrical connection, the intervening elements may include inductors and/or transformers.
Data communication: any sort of data communication scheme now known or to be developed in the future, including wireless communication, wired communication and communication routes that have wireless and wired portions; data communication is not necessarily limited to: (i) direct data communication; (ii) indirect data communication; and/or (iii) data communication where the format, packetization status, medium, encryption status and/or protocol remains constant over the entire course of the data communication.
Receive/provide/send/input/output: unless otherwise explicitly specified, these words should not be taken to imply: (i) any particular degree of directness with respect to the relationship between their objects and subjects; and/or (ii) absence of intermediate components, actions and/or things interposed between their objects and subjects.
Module/Sub-Module: any set of hardware, firmware and/or software that operatively works to do some kind of function, without regard to whether the module is: (i) in a single local proximity; (ii) distributed over a wide area; (ii) in a single proximity within a larger piece of software code; (iii) located within a single piece of software code; (iv) located in a single storage device, memory or medium; (v) mechanically connected; (vi) electrically connected; and/or (vii) connected in data communication.
Stationary exercise equipment: any exercise equipment that requires the user to remain in substantially the same location while a user uses the exercise equipment.
Non-stationary exercise equipment: any exercise equipment that allows the user to travel over a substantial distance, or real world course, while a user uses the exercise equipment, regardless of the function of the equipment with respect to the exercise; non-stationary exercise equipment includes, but is not necessarily limited to: sneakers, clothing, ice skates, bicycles (non-stationary), ankle weights, roller blades, skis, ski poles, paddle boats, row boats, pedometers, watches, rock climbing equipment, goggles, body-based exertion detection hardware; and/or body-motion-based exertion detection hardware.
Non-stationary exercise: Any exercise where the user travels over a substantial distance, or real world course.
Athletic exertion: Includes, but is not necessarily limited to human exertion associated with: running, walking, ice skating, rollerblading, skiing, swimming, rowing, pedaling, human-powered wheelchair racing and/or jogging; without limitation, athletic exertion may be measured in terms of energy, power, work, motion or the like; in some embodiments, athletic exertion may be exertion for a person's entire body (or the bodies of a group of people); in other embodiments, athletic exertion of interest may be limited to exertion of certain body parts or exertion from certain, selected body motions and/or forces.
Communications network: any type of communications network, wired, wireless or a hybrid of these types, now known, or to be developed in the future, without regard to transmission medium, nature or nomenclature of the terminal devices and/or communication protocol.
Terrain factors: any factor that tends to have a predictable affect on the amount of athletic exertion required to travel over a course; terrain factors may exist in the real world as part of a real world course; terrain factors may be virtually modeled into a virtual course; terrain factors include, but are not necessarily limited to: relative elevation, ground incline, ground conditions (for example, wet), snow condition, ice condition, altitude, wind speed, wind direction, sea condition, ground surface (for example, grass, pavement), ambient light and/or ambient temperature.
Virtual course: any abstract representation of a course, without regard to: (i) whether it corresponds to a real world course or not, (ii) how it is represented to a user (for example, schematic visual display, 2-d or 3-d graphic display, aural information, color codes, numerical readout(s)), and/or (iii) the presence or absence of user avatar(s) in the virtual course.
Virtual terrain: a virtual course with at least one terrain factor virtually modeled into it, without regard to whether the virtual terrain factor is virtually modeled in as a static future over time, or whether the terrain factor is modeled to change over time; it is noted that the absence of a terrain factor should not be considered as a virtual terrain factor, for example, a course that necessarily has a zero incline over the entire virtual course cannot be considered as virtual terrain (at least by virtue of ground inclination as the at least one terrain factor); whether something can be considered as “virtual terrain” does not depend upon: (i) whether it corresponds to a real world course or not, (ii) how it is represented to a user (for example, schematic visual display, 2-d or 3-d graphic display, aural information, color codes, numerical readout(s)), and/or (iii) the presence or absence of user avatar(s).
Normalize athletic exertion for virtual terrain: the process of taking some measurement of actual athletic exertion and mapping it to progress over a virtual terrain, where the mapping of athletic exertion to virtual progress is affected by the at least one terrain factor(s) of the virtual course; while it is generally preferable that the mapping of the athletic exertion and the affect of the virtual terrain be as realistic as feasible, this is not necessarily required for “normalizing athletic exertion for virtual terrain.”
To the extent that the definitions provided above are consistent with ordinary, plain, and accustomed meanings (as generally shown by documents such as dictionaries and/or technical lexicons), the above definitions shall be considered supplemental in nature. To the extent that the definitions provided above are inconsistent with ordinary, plain, and accustomed meanings (as generally shown by documents such as dictionaries and/or technical lexicons), the above definitions shall control.
Unless otherwise explicitly provided in the claim language, steps in method steps or process claims need only be performed in the same time order as the order the steps are recited in the claim only to the extent that impossibility or extreme feasibility problems dictate that the recited step order be used. This broad interpretation with respect to step order is to be used regardless of whether the alternative time ordering(s) of the claimed steps is particularly mentioned or discussed in this document—in other words, any step order discussed in the above specification shall be considered as required by a method claim only if the step order is explicitly set forth in the words of the method claim itself. Also, if some time ordering is explicitly set forth in a method claim, the time ordering claim language shall not be taken as an implicit limitation on whether claimed steps are immediately consecutive in time, or as an implicit limitation against intervening steps.
The present application claims priority to U.S. provisional patent application No. 61/078,100, filed on Jul. 3, 2008; all of the foregoing patent-related document(s) are hereby incorporated by reference herein in their respective entirety(ies).
Number | Date | Country | |
---|---|---|---|
61078100 | Jul 2008 | US |