This application is the U.S. national phase of PCT Application No. PCT/GB2018/051848 filed on Jun. 29, 2018, which claims priority to GB Patent Application No. 1713651.6 filed on Aug. 24, 2017, the disclosures of which are incorporated in their entirety by reference herein.
The present invention relates to a play apparatus and method.
There are times when a parent is absent from their young child, for example when they are travelling and leave the child at home with their spouse.
Parents can communicate with their pre-linguistic children when they are absent using conventional communication tools such as video conferencing.
However, such audio-visual communication only provides a limited connection between the parent and child, to the detriment of their emotional relationship.
It would be desirable for parents and children to satisfy their need to connect and enhance their emotional relationship by supporting tactile play, which provides a better connection between the parent and child (or between two users irrespective of age).
According to a first aspect of the present invention, there is provided a play apparatus comprising a pair of devices, each device comprising:
Preferably, the capture arrangement comprises a cavity.
Preferably, the cavity is configured to conceal the captured object until it is revealed by its release.
Preferably, the capture arrangement comprises a magnet.
Preferably, the capture arrangement comprises a platform.
Preferably, the release actuator is operable to release the captured object by ejecting it.
Preferably, the second device is operable to:
Preferably, the play apparatus further comprises a media output device operable to output to a user of the first device an indication of the second object having been released by the second device in response to the first release signal.
Preferably, the play apparatus further comprises a media acquisition device operable to acquire and transmit media representing the second object having been released by the second device in response to the first release signal, for output by the media output device.
Preferably, the play apparatus comprises a media output device operable to output to a user of the first device an indication of an object being captured by the second device.
Preferably, the play apparatus further comprises a media acquisition device operable to acquire and transmit media representing the other object being captured by the second device.
Preferably, the play apparatus further comprises a media sequence module operable to generate a media sequence having a duration dependent on latency from the transmission of the first release signal to the release of the second object, for output by the media output device.
Preferably, the play apparatus further comprises a media sequence module operable to generate a media sequence responsive to user input, between the detection of the capture the first object and the release of the second object, for output by the media output device.
Preferably, the media sequence module is operable to superimpose the media sequence over an output real-time media sequence representing a user of the second device.
Preferably, each device comprises an attribute detector operable to detect an attribute of an object upon its capture and wherein the transmission module of the first device is further operable to transmit a first attribute signal configured to cause the second device to apply the attribute to the second object.
Preferably, each device comprises an attribute actuator operable to apply an attribute to an object.
Preferably, applying the attribute comprises imparting a force to the object.
Alternatively, applying the attribute comprises selecting an object to release from a plurality of objects captured in the device from which the selection is made.
Preferably, the object comprises a ball.
According to a second aspect of the present invention, there is provided a method comprising the steps:
Preferably, the step of releasing the first object comprises ejecting it.
Preferably, the method further comprises the step of outputting to a user of the first device an indication of the second object having been released by the second device in response to the first release signal.
Preferably, the step of outputting to the user an indication of the second object having been released comprises acquiring and transmitting media of the second object having been released by the second device in response to the first release signal.
Preferably, the method further comprises the step of outputting to a user of the first device an indication of an object being captured by the second device.
Preferably, the step of outputting to the user an indication of an object being captured comprises acquiring and transmitting media of the other object being captured by the second device.
Preferably, the method further comprises the steps: generate and output a media sequence having a duration dependent on latency from the transmission of the first release signal to the release of the second object.
Preferably, the method further comprises the steps: generate and output a media sequence responsive to user input, between the detection of the capture the first object and the release of the second object.
Preferably, the method further comprises superimposing the media sequence over an output real-time media sequence representing a user of the second device.
Preferably, the method further comprises the step of detecting an attribute of the first object upon its capture by the first device and transmitting a first attribute signal configured to cause the second device to apply the attribute to the second object.
Preferably, the method further comprises the step of applying the attribute to the second object.
Preferably, applying the attribute comprises imparting a force to the second object.
Alternatively, applying the attribute comprises selecting an object to release as the second object from a plurality of objects captured in the second device.
Preferably, the method further comprises the step of modifying the attributes.
Preferably, the step of modifying the attributes is responsive to user input.
Preferably, the method further comprises the step of detecting an attribute of the other object upon its capture by the second device and transmitting a second attribute signal configured to cause the first device to apply the attribute to the first object.
Preferably, the method further comprises the step of applying the attribute to the first object.
Preferably, applying the attribute comprises imparting a force to the first object.
Alternatively, applying the attribute comprises selecting an object to release as the first object from a plurality of objects captured in the first device.
Preferably, the method further comprises the steps:
Preferably, the method further comprises the step of concealing the captured object in a cavity until it is revealed by its release.
Preferably, the method further comprises the step of supporting the captured object on a platform until it is released.
Preferably, the object comprises a ball.
Embodiments of the present invention will now be described, by way of example only, with reference to the drawings, in which:
Embodiments of the present invention allow a parent and child to engage in unstructured play, by providing remote, two-way, real-time, tactile “passing” of objects.
Embodiments of the present invention create the impression that a real object has been transferred from one location to another in real time. Embodiments enable dialogic communication and play without the use of language via tangible interfaces that extend the play-space beyond the screen into the physical space around it.
With reference to
In the description of
The device has a communication interface (COMM) 110 with an antenna 112, for wireless communication, such as using WiFi™, although wired communication could be used. A central processing unit (CPU) 114 and power supply (PWR) 116 are also provided. The CPU 114 controls the operation of the device and its communication. Although a CPU is convenient, the present invention is not limited to needing a CPU or other computing device. Instead other electrical circuits could be used with the detectors to trigger the sending of a release signal to the other device, or to release an object responsive to a release signal.
The memory 118 stores various program modules, which are executed by the processor to control other components of the device. These include a transmission module 120 operable to transmit a release signal, using the communication interface 110. A receiver module 122 is operable to receive a release signal, using the communication interface 110.
A media sequence module 124 is provided and its operation is described below with reference to
The device 102 has a capture arrangement, in this example a cavity 128, configured to capture (at 206 in
One or more object detector 132 is operable to detect (at 210) that the ball has been captured. The object detector 132 is connected to the CPU to provide detection signals to the CPU.
The device is operable to withhold (at 210-226) the captured ball until releasing it using one or more release actuator 134. The release actuator 134 may be used, under control of the CPU, for example to allow the ball to drop under the influence of gravity. It may be used to eject the ball, with the release actuator being all or part of a firing mechanism that fires the ball. The actuator may use DC motors or stepper motors. The release actuator may be spring-loaded, using a compression spring or a torsion spring. A motor, such as a servo may pull back the spring.
One or more attribute detector 136 is operable to detect one or more attribute of a ball upon its capture. The attribute detectors 136 are connected to the CPU to provide attribute signals to the CPU. Attributes can include motion attributes such a velocity, angle and spin. Attributes may also include visual attributes such as colour and pattern. An attribute detector may function as the object detector.
In the case where the attribute is intrinsic to the object (for example colour) the play device can identify and select an appropriate coloured object before attaching the remaining attributes to it. Thus, applying an attribute involves selecting an object to release. In that case there are two or more objects captured in the device from which the selection is made.
The transmission module 120 of the first device is further operable to control the communication interface 110 to transmit one or more first attribute signal configured to cause the second device to apply the one or more attribute to the second ball.
One or more attribute actuator 138 is operable to apply one or more attribute to an object. The attribute actuator 138 may be used, under control of the CPU, to apply a motion attribute (at 216) by imparting a force to the ball. The attribute actuator may thus be part of a firing mechanism that fires the ball with a speed, angle and/or spin that corresponds to the motion of the other ball when captured by the other device. An attribute actuator may function as the release actuator.
Other attributes may be metadata, which may be stored in the RFID tag of the ball, being read by an RFID reader as the attribute detector 136. The metadata may for example be used to identify a sound attribute, which is applied to the ball when released by the other device by playing the sound.
Attributes may be for example a surface pattern (or texture) detected on a ball (or other object). The surface texture may then be applied to the object by superimposing the identified texture on the ball's representation in a video sequence. For example, the parent may have just one ball with a marker pattern on it. The child may choose one of several balls, such as one having a particular image of its favourite comic character on its surface. The child's play device recognises the attribute of that particular image and the parent's (or child's) play device or tablet computer can augment the reality of the patterned ball at the parent's end by applying the particular image as a texture to the surface of the parent's ball (using the marker pattern for texture registration) in the video sequence of the parent in real time. The end result is that the child thinks they have passed the selected ball to the parent, whereas the parent only needs to have just one ball. The parent's tablet computer can display a notification of which ball the child passed to them, to help them talk about it with the child, if the parent doesn't notice which ball is inserted by the child into their play device. Additionally, or alternatively, an attribute may be a 3D shape of an object, which can be similarly rendered using augmented reality. Other ways of applying a surface texture include illuminating a ball from inside using LEDs or using e-ink to apply patterns to a ball or both balls.
With reference to
A first device is initially empty at 202 with no ball captured in it. A ball B is captured in the second device at 204.
A first ball A is thrown, pushed or dropped into the first device at 206. The ball B is withheld in the second device at 208. The object detector detects the capture of the ball A at 210, or if using detectors close to the hole, at 206.
The first device withholds the first ball at 210, 214, 218, 222, 226 until releasing it at 230.
Responsive to detection of the capture by the first device of the ball A, at 210 the first device transmits a first release signal 211 configured to cause the second device to release a second ball B. The release signal 211 may be an explicit command or code instructing release. The release signal 211 may be in the form of a control signal encoding attributes to be applied to the second object. Thus, the release is implicitly signalled and the release signal comprises an attribute signal. The release signal may be relayed or transformed or the release information may be transferred from one signal to another during its transmission between the devices. For example, an intermediate server may receive the release signal and generate a new release signal that is sent on to the other device. In that example, the release signal is still to be interpreted as being transmitted from one device to another.
The second device receives at 212 the first release signal 211. Responsive to the first release signal 211, the second device releases the second ball B at 216.
At this stage, the user of the first device has “passed” their ball A to the second device's user, who receives ball B. The first device at 218 withholds and conceals the captured ball A, while the second device is empty at 220.
The user of the second device can catch the ball B. The user of the first device can see that happening by use of a video conference call. The user of the second device then passes the ball back to the first device's user by throwing, pushing or dropping it into the second device at 224.
The second device detects at 224 or 228 that the other object (which may actually be the ball B or a different object) has been captured by the second device.
Responsive to the detection of the capture of the other object by the second device, the second release signal 227 is transmitted to the first device. Thus the second release signal 227 indicates that another object has been captured by the second device. In this example, the second device transmits the second release signal 227, but it may be transmitted for example by an intermediate server, which has been informed of the detection of the capture at 224 or 228.
At 230, the first device releases the ball A responsive to receipt of the second release signal 227. At 232, the second device withholds the other ball B.
This takes the play apparatus back to the initial state, thus devices at 234 and 236 are in the same state as the devices at 202 and 204.
A small child 328 is a user of a play device 324. A media output device 308, in this case a tablet computer, is positioned behind play device, from the child's point of view. The child's tablet 308 has a camera 312 with a field of view 320 that encompasses the space above the play device 324 and encompasses the child's head and shoulders 328.
The play device 324 and the child's tablet 308 are connected wirelessly to a router 302, which connects via the internet 304 to a remote router 306 for communication with the parent 330. The remote router 306 is connected to the parent's media output device 310, in this case a tablet, and their play device 326. The parent's tablet 310 has a camera 314 with a field of view 322 that encompasses the space above the play device 326 and encompasses the parent's head and shoulders 330. Tablet 310 outputs a video stream 318 acquired by the camera 312.
Tablet 308, is operable to output to a user of the first device 328 an indication 316 of the second object B having been released by the second device 326 in response to the first release signal (211 in
A media acquisition device, in this example tablet 310 with camera 314, is operable to acquire and transmit media representing the second object B having been released by the second device 326 in response to the first release signal, for output by the media output device 308. In this example, the indication is media being a live video (with sound) stream of the ball B. Another type of media is sound. The media acquisition device 314 is operable to acquire and transmit media representing the other object being captured by the second device. Another type of media is animated holographic imagery.
In this example, the tablets 308 and 310 each operate both as media output devices and media acquisition devices. At a given end, the media output device and media acquisition device may be separate devices. At a given end, the media output device and media acquisition device may be integrated in the play device.
The media sequence module (124 in
The media sequence module is operable to superimpose the media sequence 404 over an output real-time media sequence 406 representing a user of the second device.
Once the latency period has ended and the animation is complete, the media sequence 414 is rendered with a hill 418 that is empty 420. Meanwhile the media output device 412 shows 416 the ball B having been released by the second device with the media sequence 414 superimposed.
A game is instigated when two balls are placed into the paired devices.
As an example of a fetch game, the following sequence of events may be performed.
With reference to
One or more of the devices may be provided with user input components, such as buttons, joysticks, pressure and/or velocity sensors, or microphones. These components can be used for game controllers and/or attribute detectors.
The media sequence is an animation of a computer-generated virtual play space, in this example a hill 508 with a virtual ball 510 being fired away from the viewpoint by in this example a catapult 512. The catapult 512 may be controlled by user input for example a pressure/velocity sensor.
The media sequence module is operable to superimpose the media sequence 504 over an output real-time media sequence 506 representing a user of the second device.
Once the game has ended and the animation is complete, the media sequence 516 is rendered with a virtual play space, in this case a hill 520 that has an empty launch device—in this case a catapult 522. Meanwhile the media output device 514 shows 518 the real ball B having been released by the second device with the media sequence 516 superimposed.
On-screen games and associated graphics enhance the play. The users' faces are visible, with only the bottom third (approximately) used for play.
The game may be played using peripherals to control the rendered ball, such as a pressure, velocity or tilt sensitive pad, or such as a breath-sensitive “blowstick” as described below.
A blowstick can be used to blow at the screen or over the hole in the device, to control an object on a rendered surface. A blowstick can be used to blow and control a musical instrument in a rendered band.
A tilt pad can be used for example to: alter a surface rendered on screen to adjust the roll of a ball or other object, for example to adjust the path of a boat depicted on the high seas, or to adjust the path of a plane around the sky.
The method has the following steps:
602: Capture 206 a first object by a first device. The object is withheld and may be concealed.
604: Detect 206 that the first object has been captured by the first device. Attributes of the first object are detected upon its capture by the first device.
606: Modifying the attributes. The modification may be responsive to user input. For example, the user may set the direction and speed of ejection of the other object upon its release, in a catapult game as described with reference to
608: Responsive to the detection, transmit 210 a first release signal 211 from the first device configured to cause a second device to release 216 a second object (at step 612). A first attribute signal is transmitted configured to cause the second device to apply the attribute to the second object (at step 612).
610: Generate and output a media sequence (images and/or sound) having a duration dependent on latency from the transmission of the first release signal 211 to the release 216 of the second object. Additionally or alternatively, a media sequence is generated and output responsive to user input (e.g. a game), between the detection of 206 the capture the first object and the release 216 of the second object. The media sequence is superimposed over an output real-time media sequence representing a user of the second device.
612: The first release signal 211 is received 212 at the second device. Responsive to the first release signal, the second object is released 216 by the second device. The attributes detected in step 604 and 610 are applied to the second object. This may involve imparting a force to the second object.
614: Acquiring and transmitting media of the second object having been released by the second device in response to the first release signal 211. The media are output to a user of the first device as an indication of the second object B having been released 216 by the second device in response to the first release signal 211.
616: Detect that another object (which may be the second object or a different object) has been captured 224 by the second device. Responsive to the detection of the capture 224 of the other object by the second device, the second release signal 227 is transmitted (from the second device or an intermediate server) to the first device. Media of the other object being captured by the second device are acquired and transmitted. An indication of an object (ball B or a different object) being captured 224 by the second device is output to a user of the first device.
An attribute of the other object may be detected upon its capture by the second device and one or more second attribute signal may be transmitted configured to cause the first device to apply the one or more attribute to the first object. Attributes may also be modified in a step equivalent to that at 606. Applying the attribute(s) to the first object may involve imparting a force to the first object.
The first object, which has been withheld 210-226 by the first device, is released 230 responsive to receipt of the second release signal 227 configured to indicate that the other object has been captured 224 by the second device.
In the steps above, a captured object may be concealed in the cavity until it is revealed by its release. The object may comprise a ball.
A play device 702 is shown in cross section. The hole 704 has a tapered throat, to aid with catching the ball 710. It has tapered sidewalls 706, 708. A stepper motor 712 connected via its shaft 714 to an arm 716. The motor and arm assembly act as a release actuator. In this example, the releasing action is an ejecting action. As the motor turns, the end of the arm moves upwards and ejects the ball. The motor can move slowly at first to move the ball around the bend in the sidewalls that has been acting to conceal the captured ball. This reduces energy being lost to the sidewalls. Once past the bend, the motor speeds up and it applies more force to the ball, which is ejected out of the throat 704.
In the remaining Figures, features with reference numerals the same as shown in earlier Figures are the same features, and the description of those features given above in relation to the earlier Figures should be used to interpret the latter features.
In embodiments, the captured object may be supported on a platform until it is released. The platform may be on top of the play device, or within it, such as in a cavity. For example, there may be no cavity and the object may rest on the platform of the top surface of the play device, held in place by gravity.
Instead of the object detector being a sensor 132 in the play device as shown in
The object detector 932 may be simply processing logic that receives an object detection message, for example from an external object detection sensor, and causes the device to respond to detection of capture of an object.
Having three or more play devices allows users to pass balls in different ways rather than backwards and forwards. They can choose who they pass the ball to. In this example, three players can pass a ball to one or other of two players. Each player may be in a different location and has a play device. The play devices are all connected via the internet 304. Shown here, the user of the third play device 1014 puts the ball A into their right-had hole 1018. By choosing the right-hand hole, sensors associated with that hole detect the attribute, which is labelled “pass to the right”. This causes the release signal to be sent from play device 1014 to the second play device 1008, rather than play device 1002. This causes the second play device 1008 to release and eject the ball B. If the user of the third play device had put the back in the left-hand hole 1016, then the release signal would have been sent instead to the first play device 1002.
One player may act as a “games master” and they may have control of the game play of the other players. This may involve for example modifying, interrupting or overriding the release signals the release signals between the other players. Thus, they may for example block a ball pass between the other payers or bat a ball back to a player, or steal the ball to their own play device, instead of allowing it to pass to another player. The games master may act to enforce rules of game play or may suspend or terminate play. The role of games master between two or more players may be performed by a person without their own play device, or it may be performed automatically by software running on a processor in a play device or externally.
A pair of play devices 1202, 1208 are configured to act as toy car parking garages. The play devices 1202, 1208 are connected via the internet 304. The play devices may operate in the same way as described with reference to
The object may be a volume of solid, such as a ball, or a volume of liquid or gas.
Advantages of embodiments of the present invention include:
Communication is via tangible objects rather than words. Play is unstructured and mimics play with real objects when participants are in the same room. Play is in real time. Embodiments enable parents to engage with their children even when they are absent.
Number | Date | Country | Kind |
---|---|---|---|
1713651 | Aug 2017 | GB | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/GB2018/051848 | 6/29/2018 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2019/038512 | 2/28/2019 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
4844475 | Saffer | Jul 1989 | A |
4995374 | Black | Feb 1991 | A |
5041044 | Weinreich | Aug 1991 | A |
5397133 | Penzias | Mar 1995 | A |
6009458 | Hawkins et al. | Dec 1999 | A |
6359549 | Lau | Mar 2002 | B1 |
6772745 | McEachen | Aug 2004 | B2 |
9144746 | Cannon | Sep 2015 | B2 |
9320960 | Ward | Apr 2016 | B1 |
20060154711 | Ellis | Jul 2006 | A1 |
20080211771 | Richardson | Sep 2008 | A1 |
20130120364 | Aldridge et al. | May 2013 | A1 |
20130228138 | Hamill | Sep 2013 | A1 |
20140274370 | Shah | Sep 2014 | A1 |
Number | Date | Country |
---|---|---|
2010042085 | Feb 2010 | JP |
2016172815 | Nov 2016 | WO |
Entry |
---|
ISR for PCT/GB2018/051848, dated Oct. 15, 2018, 3 pages. |
Number | Date | Country | |
---|---|---|---|
20200360829 A1 | Nov 2020 | US |