System for encouraging a user to perform substantial physical activity

Information

  • Patent Grant
  • 7980997
  • Patent Number
    7,980,997
  • Date Filed
    Thursday, October 23, 2008
    16 years ago
  • Date Issued
    Tuesday, July 19, 2011
    13 years ago
Abstract
A system for encouraging a user to perform substantial physical activity. The system may include sensors that may be worn by the user while the user is performing a substantial physical activity, such as running or playing basketball. The sensors may detect the magnitude of the physical activity and may transmit data regarding the physical activity to a processing system. The processing system may display a reward to encourage the user for participating in physical activity and the reward provided may be based on the physical activity of the user.
Description
FIELD OF THE INVENTION

This application relates to an interactive system that encourages users to partake in substantial physical exercise.


DESCRIPTION OF RELATED ART

Childhood obesity in America is on the rise. Between 5-25 percent of children and teenagers in the United States are obese (Dietz, 1983). As with adults, the prevalence of obesity in the young varies by ethnic group. It is estimated that 5-7 percent of White and Black children are obese, while 12 percent of Hispanic boys and 19 percent of Hispanic girls are obese (Office of Maternal and Child Health, 1989).


Obesity presents numerous problems for the child. In addition to increasing the risk of obesity in adulthood, childhood obesity is the leading cause of pediatric hypertension, is associated with Type II diabetes mellitus, increases the risk of coronary heart disease, increases stress on the weight-bearing joints, lowers self-esteem, and affects relationships with peers. These problems are compounded by the social and psychological problems faced by children as a consequence of childhood obesity.


The three main identified causes for childhood obesity are family, low-energy expenditure and heredity. While causes such as family and hereditary require long term commitments and research, an increase in energy expenditure in children as well as adults may achieve almost immediate positive results in combating obesity.


To accomplish increased physical activity, and thereby combat obesity, the following methods of intervention treatment have been identified as considerably valuable in combating obesity, regardless of the cause; Physical Activity, Diet Management and Behavior Modification.


Physical activity, through a formal exercise program, or simply becoming more active, is valuable for burning fat, increasing energy expenditure, and maintaining lost weight. Most studies of children have not shown exercise to be a successful strategy for weight loss unless coupled with another intervention, such as nutrition education or behavior modification (Wolf et al., 1985). However, exercise has additional health benefits. Even when children's body weight and fatness did not change following 50 minutes of aerobic exercise three times per week, blood lipid profiles and blood pressure did improve (Becque, Katch, Rocchini, Marks, & Moorehead, 1988).


Many behavioral strategies used with adults have been successfully applied to children and adolescents: self-monitoring and recording food intake and physical activity, slowing the rate of eating, limiting the time and place of eating, and using rewards and incentives for desirable behaviors. Particularly effective are behaviorally based treatments that include parents (Epstein et al., 1987). Graves, Meyers, and Clark (1988) used problem-solving exercises in a parent-child behavioral program and found children in the problem-solving group, but not those in the behavioral treatment-only group, significantly reduced percent overweight and maintained reduced weight for six months.


Some systems such as the Nintendo Wii™ allow the user to expend more energy than playing sedentary computer games. However the energy used when playing these games is not of high enough intensity to contribute towards the recommended daily amount of exercise in children (BBC, 2007). Nintendo's latest iteration of an Exergame, the Wii-Fit™, provides 40 different activities; however none of them involve any outdoor activity and still require the user to be located in front of a television in order to play the game. The Exergame system requires an initial investment of hundreds of dollars for a console and the game.


Other systems that help joggers and runner's capture their physical exercise activity are only limited to capturing exercise metrics from running. Systems such as Nike Plus™ also only target users who are already health conscious and are engaging in physical activity, and only need a visualization tool to help keep track of their own user defined goals. None of the systems in the above category is tasked at educating and encouraging users to undergo substantial physical exercise, and at the same time keep them engaged.


Therefore a need exists for a system targeted towards addressing obesity, and childhood obesity in particular, using a medium that is successful with children and teenagers.


SUMMARY OF THE INVENTION

A system for encouraging a user to perform substantial physical activity may comprise one or more sensors that are configured to be worn by the user while the user is performing the physical activity. The one or more sensors may be configured to detect the magnitude of the physical activity, including movement of the user in one or more directions. The system may also comprise a user interface that is configured to provide a reward to the user for performing a substantial physical activity, other than a report about the physical activity. The system may further comprise a processing system configured to cause the user interface to provide the reward to the user based on the magnitude of the physical activity as detected by the one or more sensors. The reward generated by the user interface may be configured to display an animated game comprising an animated character, and the actions of the animated character may be correlated to the physical activity of the user.


These, as well as other components, steps, features, objects, benefits, and advantages, will now become clear from a review of the following detailed description of illustrative embodiments, the accompanying drawings, and the claims.





BRIEF DESCRIPTION OF THE DRAWINGS

The drawings disclose illustrative embodiments. They do not set forth all embodiments. Other embodiments may be used in addition or instead. Details that may be apparent or unnecessary may be omitted to save space or for more effective illustration. Conversely, some embodiments may be practiced without all of the details that are disclosed. When the same numeral appears in different drawings, it is intended to refer to the same or like components or steps.



FIG. 1 illustrates a block diagram of a system for encouraging a user to perform substantial physical activity.



FIG. 2 illustrates a detailed block diagram of the sensor module the system of FIG. 1.



FIG. 3 illustrates a block diagram of a sensor module with an onboard processor and a wired and/or wireless communication interface.



FIG. 4 illustrates a block diagram of a sensor with a wired and/or wireless communication interface and on board storage.



FIG. 5 illustrates system for encouraging a user to perform substantial physical activity without a separate sensor module.



FIG. 6 illustrates the sensor system of FIG. 1 in use by a user not participating in substantial physical activity.



FIG. 7 illustrates the sensor system of FIG. 1 in use while the user is running.



FIG. 8 illustrates the sensor system of FIG. 1 in use while the user is riding a bicycle.



FIG. 9 illustrates the sensor system of FIG. 1 in use while the user is not participating in substantial physical activity with no reward.



FIG. 10 illustrates system of FIG. 1 in use while the user of FIG. 9 is jogging with a reward shown on the user interface.



FIG. 11 illustrates system of FIG. 1 in use while the user of FIG. 9 is playing basketball with an increased reward.





DETAILED DESCRIPTION OF THE ILLUSTRATED EMBODIMENTS

Illustrative embodiments are now discussed. Other embodiments may be used in addition or instead. Details that may be apparent or unnecessary may be omitted to save space or for a more effective presentation. Conversely, some embodiments may be practiced without all of the details that are disclosed.



FIG. 1 illustrates a system for encouraging a user to perform substantial physical activity 100. As illustrated in FIG. 1, a system for encouraging a user to perform substantial physical activity 100 may consist of a sensor module 101, a processing system 200 and a second processing system 300. The components of the system 100 may be configured to be worn or held in the hand of the user and therefore may allow the user to participate in indoor and outdoor physical activities such as sports and other substantial physical activities.


The sensor module 101 may comprise a sensor or groups of sensors 125, 130, 135. The sensor(s) 125, 130, 135 may be configured to detect the magnitude of the physical activity in form of health vectors.


As used herein, a health vector may be a quantifiable snapshot of the person's physical and biological state as determined by the data gathered by the sensors, and the information extracted by the algorithms that process that data. A health vector may contain various dimensions, of which each dimension may reveal a quantifiable aspect of a person's overall health and provide the magnitude of the physical activity of the user.


A health vector may contain the following magnitude of physical activity in the form of Calories Burnt, Distance traveled, Duration of Exercise, Duration spent outdoors and Duration spent indoors. A health vector may be easily accommodated to add more dimensions on a per need basis.


As shown in detail in FIG. 2 the sensor or group of sensors 125, 130, and 135 may include a 3 Axis Accelerometer 125, a Gyroscope 130, and a GPS sensor 135. The sensor module 101 may also include an on board microcontroller 140 and a communication interface 110 which may act as a wireless communication interface that may communicate information gathered by the sensor(s) 125, 130 and 135 to a processing system 200.


The sensor module may be in the form of a wearable device, for example, a wrist watch, pendant or bracelet.


The processing system 200 may be a device with a communication interface 220 of its own, a user interface 210 and an on board microprocessor 230. The communication interface 220 of the processing system 200 may receive the information gathered by the sensor(s) 125, 130 and 135 of the sensor module 101. As used herein, a processing system 200 may be any system capable of receiving raw data regarding the physical and/or biological state of the user.


The processing system 200 may also be capable of being held in the hand of the user and may have the ability to receive the information gathered by the sensor(s) 125, 130 and 135 of the sensor module 101. Examples of a the hand held processing system may include a cell phone, mp3 player, personal digital assistant (PDA), hand held video game or hand held computer.


The processing system 200 microprocessor 230 may run algorithms on the information gathered by the sensor(s) 125, 130 and 135 of the sensor module 101 to extract quantifiable dimensions of health vector of the user. The microprocessor 230 may also run various gesture algorithms that may identify the form of physical activity the user is performing in real time.


Examples of physical gesture recognition algorithms that may run on the physical sensor data may include various substantial physical activities including sports, such as walking, running, jumping and biking.


For example, the microprocessor 230 may run various gesture algorithms to identify that the user is running, riding a bicycle, swimming, jumping rope, playing basketball or other sports or physical activities. This real time recognition may be fed into a user interface 210 which may reward the user for participating in substantial physical activity.


The user interface 210 may be an animated game with an animated character that may respond to physical activity conducted by the user, and base the animated character's daily health on the level of physical activity of the user. If at any point in the game the user neglects physical exercise, the game 210 may respond with a negative feedback for the animated character, until ultimately the animated character may abandon the user due to lack of physical exercise. A health vector may be the standard form of information that may be consumed within the game to determine the extent of in game progress and/or rewards.


The user interface 210 may also correlate the actions of the animated character on the gesture algorithms run by the microprocessor 230. For example, if the user is jumping rope, the microprocessor will identify this activity and the user interface will generate an animated character that is also jumping rope.


Continued dedication to physical activity may be rewarded by growing an in game economy that may be used to unlock new features and enhancements for the animated characters.


Examples of rewards generated by the user interface 210 may include animated games, featuring animated characters and animated scenes; and reward points. The actions of the animated characters may be correlated to the actions of the user participating in substantial physical activity. New animated scenes and animated characters may be added by the user interface as rewards for the user participating in substantial physical activity. The animated characters and animated scenes may be deleted based on a decrease or lack of substantial user physical activity.


The system 100 may also include the ability to gather game data and statistics of the game play, and communicate that data and/or information to another processing system 300. The gathered data and/or information can then be used by the users to create visualizations and statistics of their own physical activities they have performed while playing the game, and to measure those activities.


As illustrated in FIG. 3, a sensor module 301 may include a sensor or sensor(s) 120 that may be coupled with an on board microprocessor 150 or a microcontroller 140 configured to directly act on the data fed to it by the sensor(s) 120 by running pre-defined algorithms. In this configuration, the task of calculating the health vector and real time gesture recognition may be offloaded from the microprocessor 230 on the processing system 200, and be fed directly into the user interface 210 through a wired or wireless communication interface 110.


As illustrated in FIG. 4, a sensor module 401 may include sensor(s) 120, which may be coupled to a removable storage media 160, which can store data gathered from the sensor(s) 120 and/or also store health vector and gesture recognition information, the later may be possible if the sensor(s) 120 are coupled to an on board microprocessor 150 or a sophisticated microcontroller 140. This may allow the sensor(s) 120 to have the ability to store such information for a period of unspecified time and communicate the information when needed through a wired or wireless communication interface 110, allowing for offline operation instead of real time operation of the system.



FIG. 5 illustrates a single processing system 500 for encouraging a user to participate in substantial physical activity that may be contained within the single processing system 500. This processing system 500 may contain within it a sensor or array of sensors 120, which may gather data of the physical and/or biological state of the user. The processing system 500 may store the data for later processing in a storage component 260, or use an on-board microprocessor 230 to run pre-determined algorithms, and then store in its storage component 260 the resulting health vector and gesture recognition information for later use. The algorithms running on the microprocessor 230 may be a part of a user interface 210 stored in the storage component 260 of the processing system 500, or may be a part of a separate suite within the processing system 260. The microprocessor 230 may also run various algorithms in real time. This real time recognition may be fed into the user interface 210 to affect in the generate rewards on the user interface 210 that may be based on the substantial physical activity of the user. Alternatively, the information may be fed delayed offline to the user interface 210 by accessing the information from the storage 260 component of the processing system 500. The system 500 illustrated in FIG. 5 may also include the ability to gather data and statistics of the, and communicate that data and/or information to another processing system 300. The gathered data and/or information may then be used by the users to create visualizations and statistics of their own physical activities they have performed, and to measure those activities. The user's visualizations and statistics may also be used by healthcare experts and counselors to better track the progress made by the users towards reducing obesity, and to also suggest improvements and alternate regimens, which may be programmable from within the user interface 210.



FIG. 6 illustrates the use of any one of the systems for encouraging a user's participation in significant physical activity of FIG. 1 or FIG. 3, in which the user is shown wearing the sensor module 101 and holding the processing system 200. As shown in FIG. 6, the user is stationary and the user interface generates an animated character which is also shown on the display to be stationary.



FIGS. 7 and 8 illustrate the use of any one of the systems of FIGS. 1-5, in which the user is running and the user interface generates a reward in the form of an animated character. The actions of the animated character shown on the processing are shown to be based on the user's physical movement, thereby providing a reward to the user for participation in substantial physical activity. The animated character generated by the user interface may have characteristics different from those of any other animated character used in any other system. The user interface may generate an initial life-span 600 for the animated character, which may be increase or reduced based on the physical activity of the user. The life span 600 of the animated character may be increased based on the user's increased participation in substantial physical activity or the type of physical activity of the user. The life span of the animated character may be decreased based on a decrease or lack of participation in substantial physical activity of the user. The user interface may also reward the user by generating animated objects or gifts 601 for the animated character as illustrated in FIG. 8. These object or gifts may be added or removed based on the substantial physical activity of the user. The user interface may also generate new animated games featuring new animated characters and scenes based on the substantial physical activity of the user.



FIG. 9 shows a user who has not participated in any physical activity since using the systems as shown in FIGS. 1-5, and who has not received any reward points by the processing system.



FIG. 10 shows the same user of FIG. 9 shown running and viewing a user interface which provides a reward in the form of animated character and points. FIG. 11 illustrates the user of FIGS. 9 and 10 with an increased reward point tally based on participating in more substantial physical activity.


It will be recognized by those skilled in the art that the variations of the above-described sensors may readily be manufactured with conventional techniques of the type typically used in manufacturing sensor based solutions. Furthermore it is recognized by those skilled in the art that the communication interfaces of the wired and wireless type not restricted to the ones mentioned can be easily integrated with the above described configurations. It also will be recognized by those skilled in the art that various other types of processing systems can be built and, in addition, that numerous other changes can be made in the hardware and software embodiments described herein without departing from the scope and the spirit of the disclosed subject matter.


The term “coupled” encompasses both direct and indirect coupling. For example, the term “coupled” encompasses the presence of intervening circuitry between two points that are coupled. Nothing that has been stated or illustrated is intended to cause a dedication of any component, step, feature, object, benefit, advantage, or equivalent to the public, regardless of whether it is recited in the claims. In short, the scope of protection is limited solely by the claims that now follow. That scope is intended to be as broad as is reasonably consistent with the language that is used in the claims and to encompass all structural and functional equivalents.

Claims
  • 1. A system for encouraging a user to perform substantial physical activity, including substantial body movements, comprising: one or more sensors configured to be worn by the user while the user is performing the physical activity, including the body movements, the one or more sensors configured to detect body movement of the user in one or more directions;a user interface configured to provide a visual display to the user of an animated character performing an activity of the same type as the user is performing, including activity with the same type of body movements; anda processing system containing physical motion recognition algorithms configured to determine the type of activity which the user is performing based on information provided by the one or more sensors, including the information about the body movements, and, based on this determination, to cause the visual display to display to the user the animated character performing the activity of the same type as the user is performing, including activity with the same type of body movements.
  • 2. The system of claim 1, wherein the one or more sensors is configured to attach to a part of the user's body.
  • 3. The system of claim 1, wherein the processing system is configured to be held in a hand of the user while the user performs the substantial physical activity.
  • 4. The system of claim 1, wherein the user interface generates an initial life-span for the animated character.
  • 5. The system of claim 1, wherein the user interface is configured to control the life-span of the animated character, and wherein the user interface is configured to increase the life-span of the animated character based on the physical activity of the user, and,wherein the user interface is configured to decrease the life-span of the animated character based on a prolonged absence of substantial physical activity of the user.
  • 6. The system of claim 1 wherein the physical motion recognition algorithms are configured to distinguish between physical activity which constitutes walking and running and wherein the processing system is configured to cause the visual display to display the animated character walking or running, as determined by the algorithms from the information from the one or more sensors.
  • 7. The system of claim 1 wherein the physical motion recognition algorithms are configured to distinguish between physical activity which constitutes jumping and biking and wherein the processing system is configured to cause the visual display to display the animated character jumping or biking, as determined by the algorithms from the information from the one or more sensors.
  • 8. The system of claim 1 wherein the physical motion recognition algorithms are configured to distinguish between physical activity which constitutes walking, running, jumping, and biking and wherein the processing system is configured to cause the visual display to display the animated character walking, running, jumping, or bilking, as determined by the algorithms from the information from the one or more sensors.
  • 9. The system of claim 8 wherein the processing system is configured to cause the user interface to provide a reward to the user after the user performs the substantial physical activity based on information from the one or more sensors, the reward including the appearance of an object on the display, other than a character.
  • 10. The system of claim 8 wherein the processing system is configured to cause the user interface to provide a penalty to the user after the user fails to perform substantial physical activity based on information from the one or more sensors, the penalty including the disappearance of an object on the display, other than a character.
  • 11. The system of claim 1 wherein the physical motion recognition algorithms are configured to distinguish between stationary and moving physical activity and wherein the processing system is configured to cause the visual display to display the animated character as stationary or moving, as determined by the algorithms from the information from the one or more sensors.
  • 12. A system for encouraging a user to perform substantial physical activity comprising: one or more sensors configured to be worn by the user while the user is performing the physical activity, the one or more sensors configured to detect movement of the user in one or more directions;a user interface including a visual display to the user; anda processing system configured to cause the user interface to provide a reward to the user after the user performs substantial physical activity based on information from the one or more sensors, the reward including the appearance of an animated character on the visual display which did not appear in any form on the display before the user performed the substantial physical activity.
  • 13. A system for encouraging a user to perform substantial physical activity comprising: one or more sensors configured to be worn by the user while the user is performing the physical activity, the one or more sensors configured to detect movement of the user in one or more directions;a user interface including a visual display to the user; anda processing system configured to cause the user interface to provide a reward or penalty to the user after the user performs or fails to perform substantial physical activity based on information from the one or more sensors, the reward including the appearance and the penalty including the disappearance of an object on the display, other than a character.
US Referenced Citations (43)
Number Name Date Kind
5749372 Allen et al. May 1998 A
5901961 Holland, III May 1999 A
5954512 Fruge Sep 1999 A
6024675 Kashiwaguchi Feb 2000 A
6183425 Whalen et al. Feb 2001 B1
6522266 Soehren et al. Feb 2003 B1
6705972 Takano et al. Mar 2004 B1
6749432 French et al. Jun 2004 B2
7089148 Bachmann et al. Aug 2006 B1
20020103610 Bachmann et al. Aug 2002 A1
20040167420 Song et al. Aug 2004 A1
20050221960 Miyamaru et al. Oct 2005 A1
20050272504 Eguchi et al. Dec 2005 A1
20060020177 Seo et al. Jan 2006 A1
20060025282 Redmann Feb 2006 A1
20060167387 Buchholz et al. Jul 2006 A1
20060286522 Ng-Thow-Hing Dec 2006 A1
20070003915 Templeman Jan 2007 A1
20070042868 Fisher et al. Feb 2007 A1
20070087799 Van Luchene Apr 2007 A1
20070100666 Stivoric et al. May 2007 A1
20070111767 Brown et al. May 2007 A1
20070129148 Van Luchene Jun 2007 A1
20070173705 Teller et al. Jul 2007 A1
20070208232 Kovacs Sep 2007 A1
20070225071 Van Luchene et al. Sep 2007 A1
20070238499 Wright Oct 2007 A1
20070260984 Marks et al. Nov 2007 A1
20070281765 Mullen Dec 2007 A1
20080005775 Belton et al. Jan 2008 A1
20080029769 You Feb 2008 A1
20080076496 Baerlocher et al. Mar 2008 A1
20080125289 Pryor et al. May 2008 A1
20080139307 Ueshima et al. Jun 2008 A1
20080167535 Stivoric et al. Jul 2008 A1
20080216592 Ueda et al. Sep 2008 A1
20080221487 Zohar et al. Sep 2008 A1
20080311968 Hunter Dec 2008 A1
20090024062 Einarsson Jan 2009 A1
20090047644 Mensah et al. Feb 2009 A1
20090082701 Zohar et al. Mar 2009 A1
20090137933 Lieberman et al. May 2009 A1
20090299232 Lanfermann et al. Dec 2009 A1
Related Publications (1)
Number Date Country
20100105525 A1 Apr 2010 US