The present invention generally concerns unmanned aerial vehicles, airflow manipulation, and computer vision. More particularly, the present invention concerns controlling movement of an unmanned aerial vehicle by controlling airflow in an environment capable of generating airflow, and manipulation of a camera feed from the unmanned aerial vehicle.
Unmanned aerial vehicles (UAVs), sometimes referred to as “drones,” are aerial vehicles that are either autonomous, remote-controlled by a user with a control transmitter, or some combination thereof. UAVs can sometimes include cameras that record images or videos of the physical world as seen by the field of view of the camera.
Augmented reality refers to a view of a physical, real-world environment whose elements are augmented or supplemented by computer-generated sensory input. For example, augmented reality may include the view of the physical environment with text or images adding to or replacing elements of the view of the physical environment. Augmented reality may also insert or replace sounds with computer-generated sounds.
Virtual reality refers to technologies that generate, typically via compute software, a virtual world environment whose elements have little or no relationship to any physical, real-world environment. A virtual reality experience is typically intended to replace, rather than augment or supplement, an experience of any physical reality. Virtual reality typically include entirely computer-generated graphics and sounds.
Display technologies include display screens, such as liquid crystal display (LCD) display screens or organic light emitting diode (OLED) screens. Display technologies also include projectors, such as movie projectors. Displays can be included in typical monitors or televisions, in handheld devices such as cellular phones or tablet devices, or in head-mounted displays such as goggles or glasses.
A first claimed embodiment of the present invention involves a system for movement control. The system includes an air propulsion device that generates an airflow. The system also includes a porous surface, wherein the airflow generated by the air propulsion device is output through a plurality of pores in the porous surface, wherein the airflow thereby provides lift to an unmanned vehicle, wherein at least a vector component of the lift is perpendicular to the porous surface and pushes the unmanned vehicle away from the porous surface. The system also includes a communication transceiver that receives an airflow modification signal from a controller device, wherein the air propulsion device modifies the airflow in response to receipt of the airflow modification signal at the communication transceiver.
A second claimed embodiment of the present invention concerns a system for movement control. The system includes a body, wherein the body is lifted away from a porous surface at least in part by an airflow expelled through a plurality of pores in the porous surface. The system also includes a motor coupled to the body, and a propeller coupled to the motor, wherein actuation of the motor causes the propeller to spin, thereby generating a thrust, wherein at least a vector component of the thrust is parallel to the porous surface. The system also includes a communication transceiver that receives an actuation signal from a controller device, wherein the motor is actuated in response to receipt of the actuation signal.
A third-claimed embodiment of the present invention concerns a method for movement control. The method includes transmitting an activation signal to an arena system, wherein the arena system generates an airflow in response to receipt of the activation signal, the airflow providing lift to an unmanned vehicle. The method also includes receiving a thrust input via an input interface and transmitting a thrust signal to the unmanned vehicle, wherein the unmanned vehicle actuates a motorized propeller in response to receipt of the thrust signal, the motorized propeller providing thrust to the unmanned vehicle, wherein at least a vector component of the thrust is perpendicular to at least a vector component of the lift.
An arena includes a porous surface through which an airflow is output, thereby providing lift for an unmanned aerial vehicle (UAV) to push the UAV away from the porous surface. The airflow may also provide thrust to push the UAV in a direction that is parallel to the porous surface. The UAV may include one or more propellers that can provide lift, thrust, or both to the UAV. The airflow may be modified over a duration of time to modify lift or thrust to the UAV. The airflow may be modified based on regions of the arena to modify lift or thrust in different regions of the arena. The arena may include a scoreboard to display a score that may be modified as a result of actions undertaken by the UAV. Two or more UAVs may be used to play a game via the arena.
The arena system of
The arena system of
The airflow generating layer 120 of
The airflow generated by the fans 125 at the airflow generating layer 120 is then output through pores in a surface 110 of the surface layer 115. That is, the surface 110 is a porous surface. The surface 110 is illustrated with visible holes, but in other cases, the surface 110 may be made at least partially of a naturally porous fabric or other material whose pores are not easily visible.
A left unmanned aerial vehicle (UAV) 110L and a right unmanned aerial vehicle (UAV) 110R are also illustrated in
The arena system of
The arena system of
A large arena system may also include a few fans 125 similar to the one in
The arena system of
The game illustrated in
Different implementations of such a game may use baskets or hoops instead of goals 105, much like basketball, and the UAVs 100 may have a motorized or spring-loaded claw or arm or catapult to help launch the ball 140 into such a basket or hoop. Different implementations of such a game may use endzones instead of goals 105, much like American football. Different implementations of such a game may use more than one ball 140. Different implementations of such a game may use pucks rather than balls 140. Different implementations of such a game may use more UAVs 100, such as to have two teams of UAVs, with multiple UAVs 100 on each team, to make the game more similar to team sports such as soccer, basketball, lacrosse, American football, rugby, hockey, water polo, bocce ball, curling, or some combination thereof.
Sensors may be placed in the goals 105, along the surface 110, or elsewhere in the arena system to detect when the ball 140 has entered the goal. The sensors may notify the scoreboard 145 to increment the score corresponding to the appropriate UAV 100. The sensors may include pressure sensors, motion sensors, or cameras capable of recognizing the ball 140 via image recognition processes involving the arena device 180. The camera of either UAV 100 may also detect a successful goal and notify the scoreboard 145 of the goal. The sensors may include radar sensors, sonar sensors, lidar sensors, or some combination thereof. The sensors may include laser rangefinders. The ball 140 may include a wireless emitter, and the sensors may thus be receivers able to recognize when the ball is in the goal by the signals received by the ball. The ball itself 140 may include any such sensors as well, and may notify the scoreboard 145 of a successful goal itself. The sensors may also be located away from the surface 110, such as overhead or off to the side, by being mounted on a wall or ceiling or on another UAV 100 that serves as a “referee.” The sensors may also include location sensors, such as GPS receivers, which may be particularly useful if the arena is large, or triangulation-based location sensors that triangulate a location from wireless signals emitted at different known locations. The sensors may include gyroscopes and/or accelerometers onboard each of the UAVs 100.
The wireless emitter may be a sound emitter, such as a speaker, that may emit audible sound at human hearing frequencies or inaudible sound at frequencies higher or lower than human hearing frequencies. The wireless emitter may be a different type of wireless emitter, such as a Bluetooth, wifi, cellular, light, or radio emitter.
A UAV 100 may “see” the ball 140 even when it is not in the field of view of the camera of the UAV 100 by using onboard receivers corresponding to the wireless emitter in the ball 140 to sense emissions 230 output by the ball 140 and use those to detect the location of the ball 140, for example by gauging when the signal gets stronger as the UAV 100 and/or ball 140 moves.
The arena system of
For example, the left UAV 100L may gain a point by getting the flag 220 to the left base region 210L and/or by keeping the flag 220 there for a predetermined period of time. The right UAV 100R may gain a point by getting the flag 220 to the right base region 210R and/or by keeping the flag 220 there for a predetermined period of time. UAVs 100 could score points by stealing the flag 220 from each other as well.
The UAVs 100 may “see” the flag 220 even when it is not in the field of view of their camera by using onboard receivers corresponding to the wireless emitter in the flag emitter 225 to sense the emissions 230 and use those to detect the location of the flag 220, for example by gauging when the signal gets stronger as the UAV 100 and/or flag 220 moves.
The UAVs 100 may move the flag by grabbing it with a motorized or spring-loaded claw, or grab the flag 220 magnetically, or may move the flag 220 by pushing it via the body of the UAV 100 or pulling it via a hook or claw or loop or lasso mounted on the UAV 100.
The base regions 210 may be marked on the surface 110, for example by being printed on the surface 110, by being taped onto the surface 110 by a human user, by being projected onto the surface 110 via a projector (not shown), or by being displayed on the surface 110 in an embodiment wherein the surface 110 includes a display system 1170. The base regions 210 may also include wireless emitters so that the UAVs 100 know where they are even the base regions 210 are not in the field of view of the camera of the UAV 100.
Sensors on the flag 220, along the surface 110, onboard either UAV 100, or some combination thereof may be used to detect when the flag 220 has entered or exited a base region 210. The sensors may be any of the types of sensors discussed in relation to
The arena system of
The arena system of
The arena system of
Regional airflow effects 410 may be controlled by the arena device 180, by the control transmitter 800, or some combination thereof. The regional airflow effects 410 and may more specifically be produced when the airflow generating layer 120 includes multiple fans 125, and in particular, when different fans 125 are coaxed to produce different strengths and/or directions of airflow compared to the remainder of the fans 125 or just to neighboring fans 125. For example, one fan 125 can be spun faster than its neighboring fans 125, producing increased lift 1010 over the region of the surface 110 affected by the faster fan 125 compared to the regions of the surface 110 affected by the slower neighboring fans 125.
Each UAV 100 may be made aware of locations with regional airflow effects 410 by receiving signals identifying locations of regional airflow effects 410 from the arena device 180, or by receiving emissions 230 from wireless emitters located along the surface 110 that emit at locations where regional airflow effects 410 are present. Alternately, a projector may project an image or video on to the surface 110 indicating a regional airflow effect 410; this may then be recognized by the UAV 100 via image recognition.
The UAV field of view 505 of
The camera feed may be transmitted directly from the UAV 100 to the display 900, or indirectly through the arena device 180 or control transmitter 800. The camera feed may be modified before it is displayed at the display 900 to portray an augmented reality view as illustrated in
The modifications to the camera view may be made by a computing device 1100 onboard the UAV 100, by a computing device 1100 onboard the display 900, by a computing device 1100 onboard the control transmitter 800, by the arena device 180, by a separate computing device 1100 communicatively coupled to any of these devices, or some combination thereof. The modifications to the camera view may involve various computer vision techniques implemented via hardware, software, or some combination thereof. These computer vision techniques may include edge detection, tracking, pattern recognition, character recognition, 3D segmentation, 3D modeling, counting, quantification, machine learning, face detection, logo detection, optical character recognition, barcode scanning, quick response (QR) code scanning, or some combination thereof.
The camera feed may include data from multiple cameras to provide depth perception and identify distances between the UAV 100 and objects in the UAV field of view 505. The camera feed may include data from a distance measurement system, such as a laser rangefinder, a radar device, a sonar device, or a lidar device, to provide depth perception and identify distances between the UAV 100 and objects in the UAV field of view 505. Image recognition and tracking may be based on one or more shapes, one or more colors, one or more brightness levels, one or more contrast levels, a relative size, an absolute size, one or more facial features, one or more logos, one or more barcodes, one or more QR codes, one or more reference points, or some combination thereof.
The display field of view 510 of
The UAV field of view 605 of
The display field of view 610 illustrates a modified augmented reality variant of the UAV field of view 605 of
In some embodiments, UAV 100 has main body 710 with one or more arms 740. The proximal end of arm 740 can attach to main body 710 while the distal end of arm 740 can secure motor 750. Arms 740 can be secured to main body 710 in an “X” configuration, an “H” configuration, a “T” configuration, a “Y” configuration, or any other configuration as appropriate. The number of motors 750 can vary, for example there can be three motors 750 (e.g., a “tricopter”), four motors 750 (e.g., a “quadcopter”), eight motors (e.g., an “octocopter”), etc.
In some embodiments, each motor 755 rotates (i.e., the drive shaft of motor 755 spins) about parallel axes. For example, the thrust provided by all propellers 755 can be in the Z direction. Alternatively, a motor 755 can rotate about an axis that is perpendicular (or any angle that is not parallel) to the axis of rotation of another motor 755. For example, two motors 755 can be oriented to provide thrust in the Z direction (e.g., to be used in takeoff and landing) while two motors 755 can be oriented to provide thrust in the X direction (e.g., for normal flight). In some embodiments, UAV 100 can dynamically adjust the orientation of one or more of its motors 750 for vectored thrust.
In some embodiments, the rotation of motors 750 can be configured to create or minimize gyroscopic forces. For example, if there are an even number of motors 750, then half of the motors can be configured to rotate counter-clockwise while the other half can be configured to rotate clockwise. Alternating the placement of clockwise and counter-clockwise motors can increase stability and enable UAV 100 to rotate about the z-axis by providing more power to one set of motors 750 (e.g., those that rotate clockwise) while providing less power to the remaining motors (e.g., those that rotate counter-clockwise).
Motors 750 can be any combination of electric motors, internal combustion engines, turbines, rockets, etc. In some embodiments, a single motor 750 can drive multiple thrust components (e.g., propellers 755) on different parts of UAV 100 using chains, cables, gear assemblies, hydraulics, tubing (e.g., to guide an exhaust stream used for thrust), etc. to transfer the power.
In some embodiments, motor 750 is a brushless motor and can be connected to electronic speed controller X45. Electronic speed controller 745 can determine the orientation of magnets attached to a drive shaft within motor 750 and, based on the orientation, power electromagnets within motor 750. For example, electronic speed controller 745 can have three wires connected to motor 750, and electronic speed controller 745 can provide three phases of power to the electromagnets to spin the drive shaft in motor 750. Electronic speed controller 745 can determine the orientation of the drive shaft based on back-end on the wires or by directly sensing to position of the drive shaft.
Transceiver 765 can receive control signals from a control unit (e.g., a handheld control transmitter, a server, etc.). Transceiver 765 can receive the control signals directly from the control unit or through a network (e.g., a satellite, cellular, mesh, etc.). The control signals can be encrypted. In some embodiments, the control signals include multiple channels of data (e.g., “pitch,” “yaw,” “roll,” “throttle,” and auxiliary channels). The channels can be encoded using pulse-width-modulation or can be digital signals. In some embodiments, the control signals are received over TC/IP or similar networking stack.
In some embodiments, transceiver 765 can also transmit data to a control unit. Transceiver 765 can communicate with the control unit using lasers, light, ultrasonic, infra-red, Bluetooth, 802.11x, or similar communication methods, including a combination of methods. Transceiver can communicate with multiple control units at a time.
Position sensor 735 can include an inertial measurement unit for determining the acceleration and/or the angular rate of UAV 100, a GPS receiver for determining the geolocation and altitude of UAV 100, a magnetometer for determining the surrounding magnetic fields of UAV 100 (for informing the heading and orientation of UAV 100), a barometer for determining the altitude of UAV 100, etc. Position sensor 735 can include a land-speed sensor, an air-speed sensor, a celestial navigation sensor, etc.
UAV 100 can have one or more environmental awareness sensors. These sensors can use sonar, LiDAR, stereoscopic imaging, computer vision, etc. to detect obstacles and determine the nearby environment. For example, a collision avoidance system can use environmental awareness sensors to determine how far away an obstacle is and, if necessary, change course.
Position sensor 735 and environmental awareness sensors can all be one unit or a collection of units. In some embodiments, some features of position sensor 735 and/or the environmental awareness sensors are embedded within flight controller 730.
In some embodiments, an environmental awareness system can take inputs from position sensors 735, environmental awareness sensors, databases (e.g., a predefined mapping of a region) to determine the location of UAV 100, obstacles, and pathways. In some embodiments, this environmental awareness system is located entirely on UAV 100, alternatively, some data processing can be performed external to UAV 100.
Camera 705 can include an image sensor (e.g., a CCD sensor, a CMOS sensor, etc.), a lens system, a processor, etc. The lens system can include multiple movable lenses that can be adjusted to manipulate the focal length and/or field of view (i.e., zoom) of the lens system. In some embodiments, camera 705 is part of a camera system which includes multiple cameras 705. For example, two cameras 705 can be used for stereoscopic imaging (e.g., for first person video, augmented reality, etc.). Another example includes one camera 705 that is optimized for detecting hue and saturation information and a second camera 705 that is optimized for detecting intensity information. In some embodiments, camera 705 optimized for low latency is used for control systems while a camera 705 optimized for quality is used for recording a video (e.g., a cinematic video). Camera 705 can be a visual light camera, an infrared camera, a depth camera, etc.
A gimbal and dampeners can help stabilize camera 705 and remove erratic rotations and translations of UAV 100. For example, a three-axis gimbal can have three stepper motors that are positioned based on a gyroscope reading in order to prevent erratic spinning and/or keep camera 705 level with the ground. Alternatively, image stabilization can be performed digitally using a combination of motion flow vectors from image processing and data from inertial sensors such as accelerometers and gyros.
Video processor 725 can process a video signal from camera 705. For example video process 725 can enhance the image of the video signal, down-sample or up-sample the resolution of the video signal, add audio (captured by a microphone) to the video signal, overlay information (e.g., flight data from flight controller 730 and/or position sensor), convert the signal between forms or formats, etc.
Video transmitter 720 can receive a video signal from video processor 725 and transmit it using an attached antenna. The antenna can be a cloverleaf antenna or a linear antenna. In some embodiments, video transmitter 720 uses a different frequency or band than transceiver 765. In some embodiments, video transmitter 720 and transceiver 765 are part of a single transceiver.
Battery 770 can supply power to the components of UAV 100. A battery elimination circuit can convert the voltage from battery 770 to a desired voltage (e.g., convert 12v from battery 770 to 5v for flight controller 730). A battery elimination circuit can also filter the power in order to minimize noise in the power lines (e.g., to prevent interference in transceiver 765 and transceiver 720). Electronic speed controller 745 can contain a battery elimination circuit. For example, battery 770 can supply 12 volts to electronic speed controller 745 which can then provide 5 volts to flight controller 730. In some embodiments, a power distribution board can allow each electronic speed controller (and other devices) to connect directly to the battery.
In some embodiments, battery 770 is a multi-cell (e.g., 2S, 3S, 4S, etc.) lithium polymer battery. Battery 770 can also be a lithium-ion, lead-acid, nickel-cadmium, or alkaline battery. Other battery types and variants can be used as known in the art. Additional or alternative to battery 770, other energy sources can be used. For example, UAV 100 can use solar panels, wireless or inductive power transfer, a tethered power cable (e.g., from a ground station or another UAV 100), etc. In some embodiments, the other energy source can be utilized to charge battery 770 while in flight or on the ground.
Battery 770 can be securely mounted to main body 710. Alternatively, battery 770 can have a release mechanism. In some embodiments, battery 770 can be automatically replaced. For example, UAV 100 can land on a docking station and the docking station can automatically remove a discharged battery 770 and insert a charged battery 770. In some embodiments, UAV 100 can pass through docking station and replace battery 770 without stopping.
Battery 770 can include a temperature sensor for overload prevention. For example, when charging, the rate of charge can be thermally limited (the rate will decrease if the temperature exceeds a certain threshold). Similarly, the power delivery at electronic speed controllers 745 can be thermally limited—providing less power when the temperature exceeds a certain threshold. Battery 770 can include a charging and voltage protection circuit to safely charge battery 770 and prevent its voltage from going above or below a certain range.
UAV 100 can include a location transponder. For example, in a racing environment, race officials can track UAV 100 using location transponder. The actual location (e.g., X, Y, and Z) can be tracked using triangulation of the transponder. In some embodiments, gates or sensors in a track can determine if the location transponder has passed by or through the sensor or gate.
Flight controller 730 can communicate with electronic speed controller 745, battery 770, transceiver 765, video processor 725, position sensor 735, and/or any other component of UAV 100. In some embodiments, flight controller 730 can receive various inputs (including historical data) and calculate current flight characteristics. Flight characteristics can include an actual or predicted position, orientation, velocity, angular momentum, acceleration, battery capacity, temperature, etc. of UAV 100. Flight controller 730 can then take the control signals from transceiver 765 and calculate target flight characteristics. For example, target flight characteristics might include “rotate x degrees” or “go to this GPS location”. Flight controller 730 can calculate response characteristics of UAV 100. Response characteristics can include how electronic speed controller 745, motor 750, propeller 755, etc. respond, or are expected to respond, to control signals from flight controller 730. Response characteristics can include an expectation for how UAV 100 as a system will respond to control signals from flight controller 730. For example, response characteristics can include a determination that one motor 750 is slightly weaker than other motors.
After calculating current flight characteristics, target flight characteristics, and response characteristics flight controller 730 can calculate optimized control signals to achieve the target flight characteristics. Various control systems can be implemented during these calculations. For example a proportional-integral-derivative (PID) can be used. In some embodiments, an open-loop control system (i.e., one that ignores current flight characteristics) can be used. In some embodiments, some of the functions of flight controller 730 are performed by a system external to UAV 100. For example, current flight characteristics can be sent to a server that returns the optimized control signals. Flight controller 730 can send the optimized control signals to electronic speed controllers 745 to control UAV 100.
In some embodiments, UAV 100 has various outputs that are not part of the flight control system. For example, UAV 100 can have a loudspeaker for communicating with people or other UAVs 100. Similarly, UAV 100 can have a flashlight or laser. The laser can be used to “tag” another UAV 100.
Though not all of the UAV components illustrated in
While this disc-shaped embodiment of the UAV 100 is only illustrated in
While the disc-shaped embodiment of the UAV 100 is illustrated with four (4) motorized propellers, it may only have one motorized propeller, particularly if the motorized propeller can be reversibly driven. On the other hand, the UAV 100 may have more than four (4) motorized propellers as well, for additional maneuverability. In yet another embodiment (not pictured), the UAV 100 may have no motorized propellers, but may have both its lift 1010 and its thrust 1020 provided entirely by the airflow generated by the airflow generating layer 120 and output through the pores in the surface 110.
The control transmitter 800 may control the airflow of the arena system as described herein, in addition to controlling the motorized propellers of the UAV 100. While the control transmitter 800 is illustrated as wireless and separate from the arena system, it should be understood that it may be wired to the arena system, such as by appearing as controls embedded in the raised arena “table” illustrated in
In particular, the UAV 100 is experiencing lift 1010 that is perpendicular to the surface 110 and away from the surface 110. This direction may also be describes as parallel and opposite to the direction of gravity. The lift 1010 may be provided by the airflow output through the pores in the surface 110, or by the motorized propellers of the UAV 100 itself, or some combination thereof. While the direction of lift 1010 is perpendicular to the surface 110 and away from the surface 110, this does not mean that the airflow output through the pores of the surface 110 necessarily moves exactly perpendicular to the surface 110—rather, only a component of the airflow's force need provide the lift 1010 as illustrated in
The UAV 100 is also experiencing thrust 1020 in a direction that is parallel to the surface 110. This direction may also be describes as perpendicular to the direction of gravity. The thrust 1020 may be provided by the airflow output through the pores in the surface 110, or by the motorized propellers of the UAV 100 itself, or some combination thereof. While the direction of thrust 1020 is parallel to the surface 110, this does not mean that the airflow output through the pores of the surface 110 necessarily moves exactly parallel to the surface 110—rather, only a component of the airflow's force need provide the thrust 1020 as illustrated in
The airflow force 1040 is illustrated at an acute angle from the surface 110. The force diagram of
The “perpendicular” or “vertical” component 1060 is perpendicular to the surface 110 and typically along the same vector direction as the gravitational force. The perpendicular component 1060 may provide lift 1010 to the UAV 100 as illustrated in
The “parallel” or “horizontal” component 1070 is parallel to the surface 110 and typically along the vector direction that is perpendicular to the gravitational force. The parallel component 1070 may provide thrust 1020 to the UAV 100 as illustrated in
The airflow force 1040 may be at an angle as illustrated in
While airflow force 1040 is illustrated at an acute angle from the surface 110, though it should be understood that any angle is possible, including a right angle or an obtuse angle.
While the force 1040 is labeled as an “airflow” force, it should be understood that this also applies to any force generated by the motorized propellers of the UAV 100. That is, a force generated by the motorized propellers of the UAV 100 may also be broken down into component forces whose vector directions are the same as the perpendicular component 1060 and the parallel component 1070 of
The components shown in
Mass storage device 1130, which may be implemented with a magnetic disk drive or an optical disk drive, is a non-volatile storage device for storing data and instructions for use by processor unit 1110. Mass storage device 1130 can store the system software for implementing embodiments of the present invention for purposes of loading that software into main memory 1110.
Portable storage device 1140 operates in conjunction with a portable non-volatile storage medium, such as a floppy disk, compact disk or Digital video disc, to input and output data and code to and from the computer system 1100 of
Input devices 1160 provide a portion of a user interface. Input devices 1160 may include an alpha-numeric keypad, such as a keyboard, for inputting alpha-numeric and other information, or a pointing device, such as a mouse, a trackball, stylus, or cursor direction keys. Additionally, the system 1100 as shown in
Display system 1170 may include a liquid crystal display (LCD), a plasma display, an organic light-emitting diode (OLED) display, an electronic ink display, a projector-based display, a holographic display, or another suitable display device. Display system 1170 receives textual and graphical information, and processes the information for output to the display device. The display system 1170 may include multiple-touch touchscreen input capabilities, such as capacitive touch detection, resistive touch detection, surface acoustic wave touch detection, or infrared touch detection. Such touchscreen input capabilities may or may not allow for variable pressure or force detection.
Peripherals 1180 may include any type of computer support device to add additional functionality to the computer system. For example, peripheral device(s) 1180 may include a modem or a router.
The components contained in the computer system 1100 of
In some cases, the computer system 1100 may be part of a multi-computer system that uses multiple computer systems 1100, each for one or more specific tasks or purposes. For example, the multi-computer system may include multiple computer systems 1100 communicatively coupled together via at least one of a personal area network (PAN), a local area network (LAN), a wireless local area network (WLAN), a municipal area network (MAN), a wide area network (WAN), or some combination thereof. The multi-computer system may further include multiple computer systems 1100 from different networks communicatively coupled together via the Internet (also known as a “distributed” system).
The foregoing detailed description of the technology has been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the technology to the precise form disclosed. Many modifications and variations are possible in light of the above teaching. The described embodiments were chosen in order to best explain the principles of the technology, its practical application, and to enable others skilled in the art to utilize the technology in various embodiments and with various modifications as are suited to the particular use contemplated. It is intended that the scope of the technology be defined by the claim.
The present application claims the priority benefit of U.S. provisional application No. 62/402,826 filed Sep. 30, 2016 and entitled “Unmanned Aerial Vehicle Movement via Environmental Airflow,” which is hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
3279863 | Zuppiger | Oct 1966 | A |
3367658 | Bayha | Feb 1968 | A |
6021646 | Burley | Feb 2000 | A |
6075924 | Will | Jun 2000 | A |
6236365 | LeBlanc et al. | May 2001 | B1 |
6254394 | Draper et al. | Jul 2001 | B1 |
7988154 | Regan, Jr. | Aug 2011 | B1 |
8025293 | Crawford | Sep 2011 | B1 |
8909391 | Peeters et al. | Dec 2014 | B1 |
9061102 | Levien | Jun 2015 | B2 |
9218316 | Bernstein et al. | Dec 2015 | B2 |
9442485 | McDermott et al. | Sep 2016 | B1 |
9605926 | Means et al. | Mar 2017 | B1 |
9632502 | Levinson et al. | Apr 2017 | B1 |
10062292 | Blomberg et al. | Aug 2018 | B2 |
10067736 | Taylor | Sep 2018 | B2 |
10137984 | Flick | Nov 2018 | B1 |
10210905 | Castleman | Feb 2019 | B2 |
10248118 | Bernstein et al. | Apr 2019 | B2 |
20030102016 | Bouchard | Jun 2003 | A1 |
20030152892 | Huang et al. | Aug 2003 | A1 |
20040008253 | Monroe | Jan 2004 | A1 |
20040115593 | Hatlestad | Jun 2004 | A1 |
20040172187 | Wiseman | Sep 2004 | A1 |
20050004723 | Duggan et al. | Jan 2005 | A1 |
20050283281 | Hartmann et al. | Dec 2005 | A1 |
20060095262 | Danielli | May 2006 | A1 |
20060169508 | Trojahn | Aug 2006 | A1 |
20070061116 | Bush | Mar 2007 | A1 |
20070102876 | Giegerich | May 2007 | A1 |
20070130599 | Monroe | Jun 2007 | A1 |
20080073839 | Nally | Mar 2008 | A1 |
20080093796 | Narus | Apr 2008 | A1 |
20080144884 | Habibi | Jun 2008 | A1 |
20080154447 | Spinelli | Jun 2008 | A1 |
20080177994 | Mayer | Jul 2008 | A1 |
20080221745 | Diamandis et al. | Sep 2008 | A1 |
20080232602 | Shearer | Sep 2008 | A1 |
20090005167 | Arrasvuori et al. | Jan 2009 | A1 |
20090076665 | Hoisington et al. | Mar 2009 | A1 |
20090087029 | Coleman et al. | Apr 2009 | A1 |
20090118896 | Gustafsson | May 2009 | A1 |
20090125163 | Duggan et al. | May 2009 | A1 |
20090187389 | Dobbins et al. | Jul 2009 | A1 |
20090265105 | Davis et al. | Oct 2009 | A1 |
20100083038 | Pierce et al. | Apr 2010 | A1 |
20100096491 | Whitelaw et al. | Apr 2010 | A1 |
20100121574 | Ariyur et al. | May 2010 | A1 |
20100228468 | D'Angelo | Sep 2010 | A1 |
20100305724 | Fry et al. | Dec 2010 | A1 |
20110106339 | Phillips | May 2011 | A1 |
20110184590 | Duggan et al. | Jul 2011 | A1 |
20110199376 | Salemane | Aug 2011 | A1 |
20110311949 | Preston et al. | Dec 2011 | A1 |
20120009845 | Schmelzer | Jan 2012 | A1 |
20120035799 | Ehrmann | Feb 2012 | A1 |
20120093320 | Flaks et al. | Apr 2012 | A1 |
20120188078 | Soles et al. | Jul 2012 | A1 |
20120212399 | Border et al. | Aug 2012 | A1 |
20120232867 | Ahrens et al. | Sep 2012 | A1 |
20130128054 | Densham et al. | May 2013 | A1 |
20130137066 | Pollak et al. | May 2013 | A1 |
20130173089 | Bernstein et al. | Jul 2013 | A1 |
20130328927 | Mount et al. | Dec 2013 | A1 |
20140244075 | Litwinowicz et al. | Aug 2014 | A1 |
20140324253 | Duggan et al. | Oct 2014 | A1 |
20150063610 | Mossner | Mar 2015 | A1 |
20150209659 | Barr | Jul 2015 | A1 |
20150248785 | Holmquist | Sep 2015 | A1 |
20150323931 | Downey et al. | Nov 2015 | A1 |
20150346722 | Herz et al. | Dec 2015 | A1 |
20150378019 | Schissler et al. | Dec 2015 | A1 |
20160035224 | Yang et al. | Feb 2016 | A1 |
20160078759 | Nerayoff et al. | Mar 2016 | A1 |
20160091894 | Zhang et al. | Mar 2016 | A1 |
20160111006 | Srivastava et al. | Apr 2016 | A1 |
20160117853 | Zhong et al. | Apr 2016 | A1 |
20160117931 | Chan et al. | Apr 2016 | A1 |
20160196754 | Surace | Jul 2016 | A1 |
20160205654 | Robinson | Jul 2016 | A1 |
20160217698 | Liu et al. | Jul 2016 | A1 |
20160253908 | Chambers et al. | Sep 2016 | A1 |
20160257001 | Blasdel | Sep 2016 | A1 |
20160284125 | Bostick et al. | Sep 2016 | A1 |
20160291593 | Hammond et al. | Oct 2016 | A1 |
20160299506 | Bruggeman et al. | Oct 2016 | A1 |
20160330601 | Srivastava | Nov 2016 | A1 |
20160358497 | Nguyen et al. | Dec 2016 | A1 |
20170036771 | Woodman et al. | Feb 2017 | A1 |
20170039859 | Hu et al. | Feb 2017 | A1 |
20170045886 | Liu et al. | Feb 2017 | A1 |
20170053169 | Cuban et al. | Feb 2017 | A1 |
20170061813 | Tao et al. | Mar 2017 | A1 |
20170069214 | Dupray et al. | Mar 2017 | A1 |
20170116723 | Aughey | Apr 2017 | A1 |
20170158353 | Schmick | Jun 2017 | A1 |
20170165575 | Ridihalgh et al. | Jun 2017 | A1 |
20170166204 | Yoo et al. | Jun 2017 | A1 |
20170168488 | Wierzynski et al. | Jun 2017 | A1 |
20170168556 | Goslin et al. | Jun 2017 | A1 |
20170173451 | Pedersen et al. | Jun 2017 | A1 |
20170182407 | Steele et al. | Jun 2017 | A1 |
20170251323 | Jo et al. | Aug 2017 | A1 |
20170295446 | Thagadur | Oct 2017 | A1 |
20170329347 | Passot et al. | Nov 2017 | A1 |
20170337826 | Moran et al. | Nov 2017 | A1 |
20170343375 | Kamhi et al. | Nov 2017 | A1 |
20170371353 | Millinger | Dec 2017 | A1 |
20170372617 | Bruno et al. | Dec 2017 | A1 |
20180322699 | Gray et al. | Jan 2018 | A1 |
20180032071 | Wieneke | Feb 2018 | A1 |
20180039262 | Fox et al. | Feb 2018 | A1 |
20180046187 | Martirosyan et al. | Feb 2018 | A1 |
20180046560 | Gillies et al. | Feb 2018 | A1 |
20180093768 | Castleman | Apr 2018 | A1 |
20180093781 | Mallinson | Apr 2018 | A1 |
20180094931 | Taylor | Apr 2018 | A1 |
20180095433 | Rico | Apr 2018 | A1 |
20180095461 | Taylor | Apr 2018 | A1 |
20180095463 | Castleman | Apr 2018 | A1 |
20180095714 | Taylor | Apr 2018 | A1 |
20180096455 | Taylor | Apr 2018 | A1 |
20180096611 | Kikuchi | Apr 2018 | A1 |
20180098052 | Black | Apr 2018 | A1 |
20180246514 | Mitomo et al. | Aug 2018 | A1 |
20180259339 | Johnson et al. | Sep 2018 | A1 |
20180321692 | Castillo-Effen et al. | Nov 2018 | A1 |
20180329413 | Charalambides et al. | Nov 2018 | A1 |
20190019329 | Eyler et al. | Jan 2019 | A1 |
20190047700 | Liu et al. | Feb 2019 | A1 |
20190075252 | Zhao et al. | Mar 2019 | A1 |
20190079722 | Taylor | Mar 2019 | A1 |
Number | Date | Country |
---|---|---|
WO 2018063594 | Apr 2018 | WO |
Entry |
---|
U.S. Appl. No. 15/394,473, Dennis Castleman, UAV Battery Form Factor and Insertion/Ejection Methodologies, filed Dec. 29, 2016. |
U.S. Appl. No. 15/394,511, Michael Taylor, Course Profiling and Sharing, filed Dec. 29, 2016. |
U.S. Appl. No. 15/394,391, Dennis D. Castleman, Remote Controlled Object Macro and Autopilot System, filed Dec. 29, 2016. |
U.S. Appl. No. 15/394,329, Javier F. Rico, Mechanical Effects by Way of Software or Real World Engagement, filed Dec. 29, 2016. |
U.S. Appl. No. 15/394,267, Megumi Kikuchi, Collision Detection and Avoidance, filed Dec. 29, 2016. |
U.S. Appl. No. 15/394,285, Michael Taylor, Steering Assist, filed Dec. 29, 2016. |
U.S. Appl. No. 15/394,313, Michael Taylor, Proximity Based Noise and Chat, filed Dec. 29, 2016. |
U.S. Appl. No. 15/393,855, Glenn Black, Translation of Physical Object Viewed by Unmanned Aerial Vehicle Into Virtual World Object, filed Dec. 29, 2016. |
U.S. Appl. No. 15/711,961, Dominic S. Mallinson, Unmanned Aerial Vehicle Movement via Environmental Interactions, filed Sep. 21, 2017. |
U.S. Appl. No. 15/393,875, Michael Taylor, UAV Positional Anchors, filed Dec. 29, 2016. |
U.S. Appl. No. 15/394,391 Office Action dated Feb. 23, 2018. |
U.S. Appl. No. 15/393,855 Office Action dated May 16, 2018. |
Fujii, Katsuya; Higuchi, Keita; Rekimoto, Jun; “Endless Flyer: A Continuous Flying Drone with Automatic Battery Replacement”, 2013 IEEE 10th International Conference on Ubiquitous Intelligence & Computing and 2013 IEEE 10th International Conference on Autonomic & Trusted Computing, pp. 216-223. |
Williams, Elliot; “Real-life Space Invaders with Drones and Lasers,” Hackaday, Sep. 19, 2016. |
PCT Application No. PCT/US2017/048064 International Search Report and Written Opinion dated Nov. 7, 2017. |
U.S. Appl. No. 15/394,313 Office Action dated Oct. 18, 2017. |
U.S. Appl. No. 15/394,391 Office Action dated Aug. 24, 2018. |
U.S. Appl. No. 15/394,329 Office Action dated Aug. 7, 2018. |
U.S. Appl. No. 15/394,267 Office Action dated Aug. 24, 2018. |
U.S. Appl. No. 15/394,285 Office Action dated Aug. 3, 2018. |
U.S. Appl. No. 15/393,855 Final Office Action dated Oct. 12, 2018. |
U.S. Appl. No. 15/711,961 Office Action dated Oct. 5, 2018. |
Bai, Z., Blackwell, A., Coulouris, G.; Using augmented reality to elicit pretend play for children with autism. IEEE Transactions on Visualization & Computer Graphics. May 1, 2015(1):1. |
Thon S, Serena-Allier D, Salvetat C, Lacotte F.; “Flying a dron in a museum an augmented-reality serious game in a Provence”, InDigital Heritage International Congress (DigitalHeritage), Oct. 28, 2013 (vol. 2, pp. 669-676), IEEE. (Year: 2013). |
U.S. Appl. No. 15/394,329 Final Office Action dated Feb. 25, 2019. |
U.S. Appl. No. 15/394,285 Final Office Action dated Feb. 26, 2019. |
U.S. Appl. No. 15/393,855 Office Action dated Feb. 1, 2019. |
U.S. Appl. No. 16/121,441, Michael Taylor, Proximity Based Noise and Chat, filed Sep. 4, 2018. |
PCT Application No. PCT/US2017/048064 International Preliminary Report on Patentability dated Apr. 2, 2019. |
U.S. Appl. No. 15/394,267 Final Office Action dated Apr. 19, 2019. |
U.S. Appl. No. 16/121,441 Office Action dated May 15, 2019. |
U.S. Appl. No. 15/393,855 Final Office Action dated May 17, 2019. |
Number | Date | Country | |
---|---|---|---|
20180093171 A1 | Apr 2018 | US |
Number | Date | Country | |
---|---|---|---|
62402826 | Sep 2016 | US |