Aerial vehicles are most commonly operated in outdoor spaces. An aerial vehicle operating outdoors may determine a position in three-dimensional space using a position sensor, e.g., a Global Positioning System (“GPS”) receiver that captures signals from one or more satellites or other sources, as well as an inertial measurement unit (or “IMU”), one or more altimeters, barometers, or other components. An aerial vehicle operating outdoors may further detect hazards such as fixed or moving objects using any number of onboard sensors, e.g., cameras, time-of-flight sensors, acoustic sensors, or the like. An aerial vehicle operating outdoors may also typically climb above any detected hazards by increasing altitude, in order to elevate above ground-based or airborne obstacles, which may include humans or other animals, or other stationary or moving objects, or turn away from such obstacles. Because objects that are encountered outdoors, such as structures, plant life, or others, are often separated by comparatively large distances, and because outdoor spaces are typically sufficiently large, an aerial vehicle is typically able to make wide, smooth changes in course or altitude without risking contact with any obstacles, even where the aerial vehicle is operating at or near its maximum speeds.
Operating an aerial vehicle, or drone, within indoor spaces presents a unique set of challenges for the aerial vehicle, and creates unique risks for occupants or contents of the indoor spaces. Indoor spaces are commonly occupied by humans or other animals, who may approach an operating aerial vehicle with interest or intrigue, or with a lack of awareness of safety risks that may be posed by rotating motors and propellers, and attempt to extend one or more hands, fingers or other body parts into contact with such motors or propellers. For example, because indoor spaces are commonly constrained by floors and ceilings that are separated by distances of three meters or less, and feature narrow hallways or other passageways, along with limited operating areas between floors and ceilings, an aerial vehicle that operates indoors may be physically inhibited from making turns or changes in altitude that may be necessary in order to avoid intentional or inadvertent contact with advancing humans or other animals.
As is set forth in greater detail below, the present disclosure is directed to systems, such as aerial vehicles (e.g., drones), having proximity sensors for safety. More specifically, the systems and methods of the present disclosure include proximity sensors provided in selected locations on surfaces of systems, such as serial vehicles, that are configured to detect the presence of one or more body parts of humans or other animals. Such systems may include capacitive sensors having conductive components, e.g., electrodes, that are provided in selected locations on surfaces of the systems, such as on protective grates or other features provided over or near rotating or otherwise moving equipment, such as motors or propellers. The capacitive sensors may be configured to sense levels of capacitance of such components and, upon detecting a change in any of the levels of capacitance, a system may determine whether the detected change exceeds one or more thresholds, or otherwise indicates that a body part of a human or another animal is within a proximity of the system, and execute one or more actions accordingly. Such actions may include, for example, ceasing or changing operations of one or more moving components of the system (e.g., motors or propellers), causing the system to travel to another location, or any other actions. Alternatively, such actions may include the generation of one or more visual cues or audible cues, for example, by causing one or more lights to be illuminated, or causing one or more sounds to be emitted, independently or along with other actions. Where such moving components include propellers or other bladed objects, the systems and methods of the present disclosure may minimize or eliminate the risk of contact between such objects and any humans or other animals.
Referring to
The aerial vehicle 110 includes a housing 115 and a plurality of conductive components 165-1, 165-2, 165-3, 165-4, 165-5, 165-6 applied to surfaces of the housing 115. Each of the conductive components 165-1, 165-2, 165-3, 165-4, 165-5, 165-6 may operate as a sense electrode for one or more capacitive sensing systems.
The housing 115 includes an upper section 140 and a lower section 150 that mate together at a shroud (or a side section) 125 provided around a perimeter of the aerial vehicle 110. The housing 115 may be formed by joining or mounting components of the upper section 140, components of the lower section 150 and components of the shroud 125 to one another by any number of clips, screws, bolts, snap-fit connections, or any other features.
The aerial vehicle 110 further includes one or more chambers 135 for accommodating one or more other sensors, control systems, illuminators, speakers, processors, circuitry or other components. The chambers 135 may be provided in any location with respect to the housing 115, e.g., within a cavity or an enclosure defined by the housing 115, at a perimeter of the housing 115, or in any other location, e.g., on or incorporated into the upper section 140 or the lower section 150. Alternatively, the aerial vehicle 110 may include any number of other chambers or other portions for accommodating one or more sensors, control systems, processors, circuitry or other components. Such chambers or portions may be provided in any location or orientation with respect to the housing 115 or the aerial vehicle 110 as a whole. In some implementations, components of the capacitive sensing system, such as capacitors, circuits, resistors, connectors or others, may be provided within one or more of such chambers 135, or elsewhere within the housing 115 or any other portion of the aerial vehicle 110.
The upper section 140 and the lower section 150 define a cavity or an enclosure having a plurality of propulsion motors 122 disposed therein. Each of the propulsion motors 122 may be mounted within the housing 115, e.g., to an arm or another extension, or to any other surface. The propulsion motors 122 may be powered by electricity or any other fuel source or prime mover, and each of the propulsion motors 122 is coupled to a propeller 124. The propulsion motors 122 may be capable of generating sufficient rotational speeds of the propellers 124 to provide thrust and/or lift forces to the aerial vehicle 110 and any payload engaged thereby. Moreover, in some implementations, the propellers 124 may be formed from a single-piece construction, and with rigid blades of fixed dimensions. Alternatively, in some other implementations, one or more of the propellers 124 may have a foldable construction defined by one or more hinges, such that the propellers 124 may be folded into a reduced length or size when the propellers 124 are not rotating under power, or extended to a full length or size when operating under power. Furthermore, where one or more of the propellers 124 has a foldable construction, e.g., with blades having one or more hinges, in the event that an object contacts one of the propellers 124 during operation, that propeller 124 may deflect or fold in response to contact with the object. Such hinges may be provided at or near a hub of the propeller 124, or at any other location along the blades of the propeller 124.
As is shown in
As is shown in
As is further shown in
The lower section 150 may also include a lower frame (not shown in
Moreover, openings of the lower section 150 may have any radii or diameters, e.g., a diameter of approximately one to two centimeters (1-2 cm), and concentric rings may have progressively larger radii, e.g., approximately nine to ten millimeters (9-10 mm) from an inner ring defining a central opening or one another in a radial direction. Rings defining openings of the lower section 150 may have cross-sections of any other shape or size, e.g., squares, rectangles, circles or others, and may have any dimensions, such as widths or diameters of one to one-and-one-half millimeters (1-1.5 mm) each, and any lengths.
The landing pegs 158 may be any extensions or appurtenances extending below the lower section 150 that enable the aerial vehicle 110 to complete a landing evolution on any substantially flat surface. For example, the landing pegs 158 may have lengths that extend below a lowest point of the housing 115 or any chambers extending beneath the lower section 150, and ends that are aligned in a common plane, such that the aerial vehicle 110 may land on a substantially flat landing surface with each of the ends of the landing pegs 158 in contact with the landing surface, and without requiring any other portion of the aerial vehicle 110 to contact the landing surface. The landing pegs 158 may also assist or guide the aerial vehicle 110 when landing into or onto a docking station (e.g., a dock) or other system. For example, in some implementations, a docking station may include one or more portions having holes or other openings that correspond to or may accommodate the extensions or appurtenances therein.
The landing pegs 158 may be constructed and coupled to the housing 115 in any manner. For example, in some implementations, the landing pegs 158 may be discrete components that are fastened or otherwise joined or affixed to undersides of support members of the lower section 150, or to undersides of mesh sections of the lower section 150. Alternatively, the landing pegs 158 may be formed as integral parts of the lower section 150, e.g., mesh sections, arms of chambers to which the propulsion motors 122 are joined, or to undersides of the propulsion motors 122 themselves, and extend downward therefrom along common rotational axes of shafts or other portions of the propulsion motors 122, or in any other location or orientation with respect to any aspects of the propulsion motors 122.
The shroud 125 may be a single component, or a plurality of components, provided about a perimeter of the housing 115, and which joins the upper section 140 and the lower section 150 to form the housing 115. The shroud 125 may have a shape that generally defines the shape of the cross-section of the housing 115, e.g., a rounded square, or a squircle. As is shown in
The shroud 125 further includes a pair of indentations 126 on opposite sides of the aerial vehicle 110. The indentations 126 may be sized and oriented to accommodate one or more fingers of a hand, thereby enabling a human operator to hand-carry or otherwise manually transport the aerial vehicle 110 by grasping the shroud 125 and inserting one or more fingers into each of the indentations 126. Although the aerial vehicle 110 of
As is shown in
In some implementations, the LIDAR sensor 130-1 may be configured to calculate distances to objects at an angular range of up to three hundred sixty degrees about the LIDAR sensor 130-1, and at any maximum measuring radius, e.g., twelve meters (12 m) or more. For example, in some implementations, the LIDAR sensor 130-1 may be configured to transmit light at any measurement frequency, e.g., up to 4,500 Hertz (or 4,500 measurements per second), and the transmitted light may be infrared light, or light at any other wavelength or frequency. In some implementations, the LIDAR sensor 130-1 may transmit light in a single direction, and capture or receive reflections of the light from one or more objects. In some other implementations, the LIDAR sensor 130-1 may be configured to transmit light along a first axis, e.g., normal to a yaw axis of the aerial vehicle 110, and to rotate the first axis about a second axis, e.g., along or parallel to the yaw axis of the aerial vehicle 110. The LIDAR sensor 130-1 may further combine distances calculated based on times-of-flight with information regarding angles at which the light was emitted and received to generate a point cloud or another representation of positions of objects, which may be stored by the aerial vehicle 110 and processed or transmitted to one or more external devices or systems for processing. The LIDAR sensor 130-1 may also include one or more motor drivers or other components for controlling a start, a stop or a speed of a motor, and may generate and transmit packets of data including any information regarding transmissions or returns, e.g., rotational speeds, start or end angles, time stamps, or other related data. Point clouds or other sets of distance or position information captured by the LIDAR sensor 130-1 may be utilized to construct an environment map (e.g., a physical map) of an indoor space in which the aerial vehicle 110 operates, or for any other purpose.
As is also shown in
The aerial vehicle 110 further includes another time-of-flight sensor 130-2 (or a time-of-flight sensor module) having a field of view that extends normal from a substantially flat face and is aligned at an edge of the housing 115, e.g., radially outward from the shroud 125 and within a cross-section of the housing 115. The time-of-flight sensor 130-2 may capture data and determine ranges to objects forward of the aerial vehicle 110 based on the captured data in a manner similar to that described above with respect to the time-of-flight sensor 130-2 provided in association with the upper section 140 and described above.
The aerial vehicle 110 also includes a pair of sets of illuminators 132, with each of the sets of illuminators 132 aligned in a common line. The sets of illuminators 132 may include one or more light-emitting diodes (or “LED”), incandescent bulbs, compact and/or linear fluorescent bulbs, halogen lamps, metal halide lamps, neon lamps, sodium lamps or any other type or form of lights configured to project visible light at any frequency, wavelength or intensity during operations of the aerial vehicle 110. For example, in some implementations, each of the sets of illuminators 132 may include a plurality of white lights. Alternatively, the sets of illuminators 132 may include one or more lights in colors other than white.
The aerial vehicle 110 also includes a pair of cameras 130-3 (or camera modules, or other imaging devices of any type or form). The cameras 130-3 have fields of view with axes of orientation extending normal from another substantially flat face of the chamber 135 that is angled in a downward direction with respect to the housing 115, e.g., below the cross-section of the housing 115, or otherwise aligned at a non-parallel angle with respect to the other substantially flat face of the chamber 135. Alternatively, either or both of the cameras 130-3 may be aligned with fields of view or axes of orientation extending in any direction or orientation with respect to the aerial vehicle 110, e.g., forward, aft, port, starboard, up, down, or in any other direction, or the yaw, pitch or roll axes of the aerial vehicle 110, or at any other angle or orientation. The cameras 130-3 may be aligned with fixed orientations, or with variable orientations (e.g., rotating or shifting orientations).
Each of the cameras 130-3 may be configured to capture visual imaging data or depth imaging data with respect to objects or features that are present within their respective fields of view. In some implementations, depth imaging data captured by the cameras 130-3 may be interpreted to detect objects within an indoor space, to calculate distances or ranges to such objects, or to identify boundaries, edges or other features of such objects. In some implementations, visual imaging data captured by the cameras 130-3 may likewise be processed to identify or detect one or more boundaries or other features of objects, e.g., based on one or more edges, contours, outlines, colors, textures, silhouettes, shapes or other characteristics of objects depicted within the visual imaging data.
Depth imaging data and/or visual imaging data captured by either of the cameras 130-3 may be utilized for any purpose. In some implementations, the depth imaging data and/or visual imaging data may be processed to construct an environment map (e.g., a physical map) of an indoor space, or to select courses, speeds or altitudes for avoiding objects within the indoor space. In some implementations, the depth imaging data and/or visual imaging data may be processed to determine information regarding any events or conditions that may be occurring within an indoor space, or to identify such events or conditions. In some implementations, the depth imaging data and/or visual imaging data may be captured, processed or stored by one or more components of the aerial vehicle 110. In some other implementations, however, the depth imaging data and/or visual imaging data may be transmitted to one or more external systems that may be provided within an indoor space where the aerial vehicle 110 operates, or on or in one or more alternate or virtual locations, such as in a “cloud”-based environment. Imaging data captured by either of the cameras 130-3 may be utilized for visual-inertial odometry (or “VIO”), for monitoring or evaluating conditions within an indoor space, or for any other purpose. In some implementations, one or more of the cameras 130-3 may be an electronic shutter camera, a global shutter camera, a rolling shutter camera, or any other type or form of camera. In some implementations, imaging data captured by one or more of the cameras 130-3 may be continuously streamed to one or more external devices or systems, or processed and stored for any purpose.
Furthermore, in some implementations, the cameras 130-3 may have fields of view that are generally oriented in a common direction and substantially overlap. In some implementations, one of the cameras 130-3 may be a visual camera (e.g., a color camera) that may be used to capture images that are encoded as video files and transmitted to a remote system for streaming to a user device. In some other implementations, one of the cameras 130-3 may be a global shutter visual graphics array (or “VGA”) camera (e.g., a grayscale camera) having an infrared-cut filter, e.g., at 650 nanometers (or 650 nm), and a field of view of 166 degrees diagonal, 133 degrees horizontal, and 100 degrees vertical. Where one of the cameras 130-3 is a global shutter camera, the one of the cameras 130-3 may be configured to capture image frames at a rate of approximately thirty frames per second (or 30 fps), and such images may be used by the aerial vehicle 110 for visual simultaneous localization and mapping (or “SLAM”), or for any other purpose.
In addition to the LIDAR sensor 130-1, the time-of-flight sensors 130-2 and the cameras 130-3, the aerial vehicle 110 may further include any type or form of other sensors, or any number of such sensors. For example, in some implementations, the aerial vehicle 110 may include one or more speakers or ultrasonic sensors (or an ultrasonic sensor module) that are configured to transmit sound waves or pulses within an indoor space, and reflections of sound waves or pulses transmitted by the speakers or ultrasonic sensors within the indoor space may be captured and interpreted to calculate distances or directions from the aerial vehicle 110 to one or more objects within the indoor space based on times at which the sound waves or pulses were transmitted and received, and the speed of sound within the indoor space.
The various components of the housing 115, including but not limited to the upper section 140, the lower section 150, the chambers 135, the landing pegs 158 or others may be formed from any suitable materials, such as plastics (e.g., thermosetting plastics such as epoxy or phenolic resins, polyurethanes or polyesters, as well as polyethylenes, polypropylenes or polyvinyl chlorides, and also recycled plastics), wood (e.g., woods with sufficient strength properties such as ash), metals (e.g., lightweight metals such as aluminum, or metals of heavier weights including alloys of steel), composites (e.g., carbon fiber), rubbers, or any other combinations of materials.
For example, in some embodiments, the upper section 140, the lower section 150, the chambers 135, the landing pegs 158 or others may be formed from polycarbonates, acrylonitrile butadiene styrene (or “ABS”), a blend of polycarbonates and ABS, or any other suitable materials, e.g., plastic, wood, metal, composites, carbon fibers, rubbers, or any other suitable materials. The upper section 140, the lower section 150, the chambers 135, the landing pegs 158 or others may be further formed from any number of portions, pieces or sections, such as one, two (viz., the upper section 140 or the lower section 150), or more.
The aerial vehicle 110 may further include any number of processors, memory components or data stores, transceivers, or other processor-driven equipment (not shown), which may be maintained or stored within the housing 115, the chambers 135 or any other portion of the aerial vehicle 110. The aerial vehicle 110 may be configured for communication via any number of wireless systems or protocols, including but not limited to wireless fidelity (“Wi-Fi”), Bluetooth, radio frequency identification (or “RFID”), near-field communication (or “NFC”) readers, or any other type of systems or protocols.
The conductive components 165-1, 165-2, 165-3, 165-4, 165-5, 165-6 may have any size, shape or dimensions (e.g., lengths, widths or thicknesses), and may be formed from any suitable materials, including but not limited to copper, aluminum or any other conductive or semi-conductive materials. The conductive components 165-1, 165-2, 165-3, 165-4, 165-5, 165-6 may be applied to surfaces of the housing 115 in any manner, such as by one or more fasteners, such as clips, screws, bolts, snap-fit connectors, or any other like features, or one or more adhesives, such as glues, tapes or other materials.
For example, sizes, shapes or dimensions of the conductive components 165-1, 165-2, 165-3, 165-4, 165-5, 165-6 may be selected based on sizes, shapes or dimensions of portions of the housing 115 to which such components are joined, or on any other factors. In some implementations, a size, a shape, or a dimension of a conductive component may be selected to increase or decrease sensitivity of a capacitive sensing system, which may be defined or determined based on the size, the shape or the dimension of the conductive component, dielectric properties of the portions of the housing 115 to which the conductive component is joined, or any other factors. In some implementations, the portions of the housing 115 to which the conductive components 165-1, 165-2, 165-3, 165-4, 165-5, 165-6 are joined may be formed from any suitable materials, such as plastics, polyethylenes, woods or other features having comparatively high dielectric contents to aid in propagating fields, and to avoid reducing a field strength accordingly. In some implementations, such portions of the housing 115 may be formed in any manner, such as by injection molding around one or more of the conductive components 165-1, 165-2, 165-3, 165-4, 165-5, 165-6.
The conductive components 165-1, 165-2, 165-3, 165-4, 165-5, 165-6 may be provided in any location with respect to rotating or otherwise moving components of the aerial vehicle 110, e.g., the motors 122 and the propellers 124. For example, in some implementations, such as is shown in
In some implementations, conductive components may be integral portions of the upper section 140 or the lower section 150. For example, in some implementations, conductive components may be included in rings or other features defining a central opening, spokes or other portions of a lattice structure provided above or near the motors 122 or the propellers 124.
Conductive components provided on surfaces of systems, e.g., the aerial vehicle 110, may be linked or coupled to one or more features of a capacitive sensing system, which may include one or more circuits, resistors, capacitors, voltage sources, connections to ground or others. The capacitive sensing system may be configured to determine levels of capacitance on the conductive components. As is shown in
In accordance with one or more implementations of the present disclosure, an integrated circuit component may be coupled to one or more conductive components that are operating as a sense electrodes. In particular, a conductive component may be coupled to a first terminal of an integrated circuit component using a first line. A capacitor may be coupled to the first line, and also to a second terminal of the integrated circuit component. The integrated circuit component may be configured to determine a change in capacitance of a system including the conductive component that was caused by introduction of a new object, e.g., a human hand or finger, nearby.
Various characterizations may be utilized for such a system. In some implementations, a conductive component may act or be characterized as a sense electrode for a capacitive touch sensor system. In such implementations, capacitance measured using a sense electrode may be characterized as “self-capacitance.” Alternatively, a conductive component may act as one half or “plate” of a capacitor system, with a ground plane representing the other half or “plate” of the capacitor system.
In some implementations of the present disclosure, movement of a finger closer to a conductive component of a capacitor system causes a change in an amount of charge that can be stored or held by the capacitor system at a given voltage level. For example, charged particles of a hand, a finger or another body part may have an impact on an electromagnetic field or photon field, which may in turn impact the amount of charge that can be stored or held by the capacitor system.
In some implementations of the present disclosure, an integrated circuit component is configured to determine a change in capacitance of a capacitor system comprising a conductive component by repeatedly charging and discharging a capacitor, and measuring an amount of time required to discharge the capacitor, e.g., to a voltage below a predetermined threshold.
In some implementations of the present disclosure, an integrated circuit component is configured to determine a change in capacitance of a capacitor system based on an output from a charging and discharging capacitor, which may be provided as an input to a first counter that counts a number of rising or falling edges during a predetermined or selected time period. If capacitance increases, e.g., based on a finger moving closer to a conductive component, a frequency of an output signal will decrease due to an increased amount of time required to discharge the capacitor, and a count of rising or falling edges may also decrease.
In some implementations of the present disclosure, a second counter may be utilized to determine whether a count threshold has been reached following a first count. If a count threshold is not reached within a set amount of time from a first count, the second counter may reset. Alternatively, an evaluation algorithm may be utilized to determine how many count events have occurred within a defined time period prior to a most recent count event. Upon determining that a count threshold has been reached, an integrated circuit component generates a signal indicating a change in capacitance above a threshold has been observed.
The systems and methods of the present disclosure may be used to detect a human or another animal within a proximity of an operating system, such as an aerial vehicle, based on changes in capacitance on one or more conductive components of a capacitive sensing system. As is shown in
As is shown in
Upon determining that one or more of the changes ΔC1, ΔC3, ΔC4 in levels of capacitance, or the rates of change of such levels exceed a predetermined threshold, the circuit 162 may generate and transmit one or more signals to a control system of the aerial vehicle 110, which may then execute one or more actions in response to such signals. For example, as is shown in
Alternatively, in some implementations, upon detecting the changes ΔC1, ΔC3, ΔC4, and determining that one or more changes exceeds a predetermined threshold, the aerial vehicle 110 may execute one or more sets of instructions to cause one or more of the motors 122 to immediately stop, e.g., to operate at a speed of zero, thereby causing the aerial vehicle 110 to descend in altitude, away from the position P1. In some embodiments, while descending, the aerial vehicle 110 may sense the change in altitude and execute one or more sets of instructions to cause the motors 122 to reengage in operation once the aerial vehicle 110 is sufficiently far from the position P1. As yet another alternative, upon detecting the changes ΔC1, ΔC3, ΔC4, and determining that one or more changes exceeds a predetermined threshold, the aerial vehicle 110 may generate a visual cue by activating one or more illuminators to project light therefrom, generate an audible cue by causing one or more sounds (such as tones, sirens or spoken words such as “stay away!”) to be emitted from one or more speakers, or take any other actions.
Accordingly, the systems and methods of the present disclosure include capacitive sensing systems having conductive components provided in selected locations on surfaces of systems, such as aerial vehicles, that are configured to detect the presence of one or more body parts of humans or other animals. Upon detecting a change in levels of capacitance of any of the conductive components, a system may determine whether the detected change exceeds one or more thresholds, or otherwise indicates that a body part of a human or another animal is within a proximity of the system, and execute one or more actions accordingly. For example, the system may halt or vary operations of one or more components of the system (e.g., motors or rotating propellers), cause the system to travel to another location, or otherwise avoid contact, or minimize a risk of contact, between components of the system and any humans or other animals.
In some implementations, aerial vehicles of the present disclosure may be programmed or configured to generate one or more environment maps of indoor spaces within a facility, based on images or other data captured by one or more sensors, e.g., the sensors 130-1, 130-2, 130-3. An environment map may identify or depict one or more boundaries (e.g., walls, ceilings, floors) or other aspects of indoor spaces, as well as the respective dimensions of such spaces, or the respective surfaces or textures of such boundaries. In some embodiments, an aerial vehicle may autonomously travel throughout an indoor space in order to capture data using one or more sensors, and such data may be utilized in generating an environment map of the indoor space. In some other embodiments, the aerial vehicle may be transported (e.g., carried) or escorted by a human actor throughout an indoor space, and may capture data using one or more sensors as the aerial vehicle is transported or escorted throughout the indoor space. Data captured as the aerial vehicle is escorted may be utilized in generating an environment map of the indoor space. Additionally, in some embodiments, the aerial vehicle may selectively operate one or more propulsion motors as the aerial vehicle is transported or otherwise escorted throughout an indoor space, in order to maintain altitude and/or tilt control. Furthermore, in some embodiments, data captured by the aerial vehicle as the aerial vehicle travels throughout the indoor space may be adjusted to account for presence of one or more body parts of a human actor that is transporting or otherwise escorting the aerial vehicle through the indoor space. Additionally, an environment map generated based on data captured using one or more sensors of an aerial vehicle may further include one or more layers identifying locations of utilities, services or other systems within an indoor space.
Aerial vehicles of the present disclosure may be configured to operate any access-related systems within an indoor space, including any portals or other systems, or to enable communication with persons within an indoor space via one or more secure communications channels (e.g., voice, video and/or data). In some embodiments, an aerial vehicle may be programmed or instructed to automatically open a door or other entry point in a home, an office or another structure (e.g., a private dwelling or business location), to access a public garage or other parking or standing area, or to activate an indication within an indoor space. For example, an aerial vehicle may be configured to transmit wireless codes, signals or other information to trigger a response from one or more devices or machines equipped with wireless transceivers, such as garage door openers, doorbells, lights, haptic feedback systems or other machines or devices. Aerial vehicles may also be configured to share any relevant information or data with persons, while preventing such persons from accessing any other sensitive information or data within the indoor space. Aerial vehicles may be further configured to communicate with any other vehicles or systems within an indoor space, including but not limited to cleaning robots, servicing robots, domestic robots, household robots, monitoring and/or surveillance robots, or robots of any other kind, e.g., to provide one or more instructions for performing one or more tasks within the indoor space.
In accordance with some embodiments of the present disclosure, an aerial vehicle may be configured to operate along with one or more docking stations, e.g., base components, charging docks (or charging stations or docking stations), or other intermediary devices. Such stations may have openings, cavities or spaces configured to accommodate one or more portions of an aerial vehicle, and may include one or more surfaces that are aligned to come into contact with corresponding surfaces of the aerial vehicle, thereby enabling electrical power, information or data to be conveyed between the aerial vehicle and such stations. In some embodiments, a base component, a charging dock, or another intermediary device may include an opening, a cavity or another space that is sized and shaped to receive or accommodate a specific portion of an aerial vehicle, e.g., a fuselage of the aerial vehicle, and to enable the aerial vehicle to be aligned in one of a plurality of alignments or orientations with respect to the base component, the charging dock or intermediary device.
For example, in some embodiments, an aerial vehicle may also include one or more contacts that are positioned or aligned in an arrangement or pattern on an external surface of a portion of the aerial vehicle, e.g., a fuselage of the aerial vehicle, that is designed to contact corresponding portions of a base component, a charging dock or another intermediary device. One or more internal surfaces of the base component, the charging dock or the intermediary device may include contacts that are positioned or aligned in an arrangement or pattern that is consistent with arrangements or patterns of contacts on external surfaces of an aerial vehicle, such that the contacts on the external surface of the portion of the aerial vehicle and the contacts on the one or more internal surfaces of the base component, the charging dock or the intermediary device mate or otherwise come into contact with one another when the portion of the aerial vehicle is in one of a plurality of predetermined alignments or orientations.
In some embodiments, an aerial vehicle may include one or more cameras or other optical sensors having fields of view that extend below the aerial vehicle and are aligned to capture imaging data while the aerial vehicle executes one or more landing evolutions. Thus, when the portion of the aerial vehicle attempts to land on a docking station or at another location, such cameras or other optical sensors may be used to guide the aerial vehicle onto the docking station or another landing surface.
Moreover, aerial vehicles of the present disclosure may be configured for use or operation within indoor spaces or facilities of any kind. As used herein, the term “facility” shall refer to any building, region, structure or other space (e.g., covered or uncovered), such as a home of any type, kind, shape or form, including but not limited to a house, an apartment, a condominium, a dormitory, a barracks, or any other defined or undefined structure having one or more living spaces. A facility may also be a business-related structure such as a building, an office, a shopping center, a restaurant, a post office, a grocery store, a department store, a materials handling facility, or any other defined or undefined structure having one or more commercial areas. A facility may also be any other type of facility including but not limited to stadiums, ballfields, transportation centers or financial institutions (e.g., banks).
In some implementations, the aerial vehicles of the present disclosure may be dispatched or otherwise instructed to travel to one or more locations, to capture images or other data in such locations, and to determine whether a predetermined event or condition (e.g., an alarm event or condition) is occurring or has occurred at the one or more locations. In some embodiments, the aerial vehicles may be dispatched or instructed to travel to such locations directly, e.g., in response to one or more sets of instructions received from an external computer device over one or more networks.
The aerial vehicles of the present disclosure may take any action in response to determining that one or more sensors has detected a predetermined event or condition, or upon processing or otherwise analyzing data captured from one or more spaces within a facility. For example, an aerial vehicle may affirmatively determine, e.g., to various levels of confidence, that a predetermined event or condition has occurred in one or more spaces within a facility or, alternatively, that the predetermined event or condition has not occurred in such spaces. Alternatively, the aerial vehicles of the present disclosure may transmit images or other data captured from such spaces to one or more external computer devices or systems, such as a base component or other station configured to charge the aerial vehicles with electrical power (e.g., a base component, a charging dock, or another intermediary device or system) within the facility or, alternatively, to a server or other computer device or system provided in one or more alternate or virtual locations, e.g., in a “cloud”-based environment. The images or other data may be transmitted independently or along with one or more measures of a level of confidence (e.g., a confidence score) as to whether a predetermined event or condition has occurred or is present within the one or more spaces. Subsequently, the external computer device or system may make a determination, based on the images or other data, or any other data captured by one or more sensors present within such spaces or in any other location, whether a predetermined event or condition has occurred or is present within such spaces, or whether the predetermined event or condition has not occurred or is not present in such spaces.
Moreover, upon determining that a predetermined event or condition has occurred or is present within one or more spaces of a facility, an aerial vehicle of the present disclosure, or a base component, a charging dock or another intermediary device in communication with the aerial vehicle, may transmit one or more requests or sets of instructions requesting additional resources at such spaces to one or more external computer devices or systems.
Referring to
The aerial vehicle 210 may be any type or form of aerial vehicle (e.g., an unmanned aerial vehicle, or drone) that may be programmed or configured to autonomously perform one or more operations within indoor spaces of a facility. As is shown in
The processor 212 may be configured to perform any type or form of computing function associated with the operation of the aerial vehicle 210. For example, the processor 212 may be configured to execute any other algorithms or techniques (e.g., machine learning systems or techniques) associated with one or more applications, purposes or functions, such as navigation, monitoring or collision avoidance, or to select at least one of a course, a speed or an altitude for the safe operation of the aerial vehicle 210. The processor 212 may be configured to control any aspects of the operation of the aerial vehicle 210 and any computer-based components thereon, including but not limited to the motors 222 or the sensors 230. For example, the processor 212 may control the operation of one or more control systems or modules, such as the control system 220, for generating instructions for conducting operations of one or more of the motors 222 or the sensors 230. Such control systems or modules may be associated with one or more other computing devices or machines, and may communicate with the data processing system 280 or one or more other computer devices over the network 290, through the sending and receiving of digital data.
The processor 212 may be a uniprocessor system including one processor, or a multiprocessor system including several processors (e.g., two, four, eight, or another suitable number), and may be capable of executing instructions. For example, in some embodiments, the processor 212 may be a general-purpose or embedded processor unit such as a CPU or a GPU having any number of instruction set architectures (ISAs), such as the x86, PowerPC, SPARC, or MIPS ISAs, or any other suitable ISA. Where the processor 212 is a multiprocessor system, each of the processors within the multiprocessor system may operate the same ISA, or different ISAs.
Additionally, the aerial vehicle 210 further includes one or more memory or storage components 214 (such as databases or data stores) for storing any type of information or data, e.g., instructions for operating the aerial vehicle 210, or information or data captured during operations of the aerial vehicle 210. The memory 214 may be configured to store executable instructions, imaging data, paths or routes, control parameters and/or other data items accessible by or to the processor 212. The memory 214 may be implemented using any suitable memory technology, such as random-access memory (or “RAM”), static RAM (or “SRAM”), synchronous dynamic RAM (or “SDRAM”), non-volatile/Flash-type memory, or any other type of memory. In some embodiments, program instructions, imaging data, flight paths, flight control parameters and/or other data items may be received or sent via the transceiver 216, e.g., by transmission media or signals, such as electrical, electromagnetic, or digital signals, which may be conveyed via a communication medium such as a wired and/or a wireless link.
The transceiver 216 may be configured to enable the aerial vehicle 210 to communicate through one or more wired or wireless means, e.g., wired technologies such as Universal Serial Bus (or “USB”) or fiber optic cable, or standard wireless protocols or standards such as Bluetooth or any Wi-Fi protocol, over the network 290 or directly. The transceiver 216 may further include or be in communication with one or more input/output (or “I/O”) interfaces, and may be configured to allow information or data to be exchanged between one or more of the components of the aerial vehicle 210, or to one or more other computer devices or systems (e.g., other aerial vehicles, not shown) via the network 290. For example, in some embodiments, the transceiver 216 may be configured to coordinate I/O traffic between the processor 212 and one or more onboard or external computer devices or components, e.g., the control system 220, or the motors 222 or sensors 230. The transceiver 216 may perform any necessary protocol, timing or other data transformations in order to convert data signals from a first format suitable for use by one component into a second format suitable for use by another component. In some embodiments, the transceiver 216 may include support for devices attached through various types of peripheral buses, e.g., variants of the Peripheral Component Interconnect (PCI) bus standard or the Universal Serial Bus (USB) standard. In some other embodiments, functions of the transceiver 216 may be split into two or more separate components, or integrated with the processor 212.
The control system 220 may include one or more electronic speed controls, power supplies, navigation systems and/or payload engagement controllers for controlling the operation of the aerial vehicle 210 and for engaging with or releasing items, as desired. For example, the control system 220 may be configured to cause or control the operation of one or more of the motors 222, e.g., to cause one or more of the motors 222 to operate at desired speeds, in order to guide the aerial vehicle 210 along a desired course, at a desired speed, or at a desired altitude, as appropriate. The control system 220 may further control any other aspects of the aerial vehicle 210, including but not limited to the operation of one or more steering or braking systems of a ground vehicle, or one or more control surfaces such as wings, rudders, flaperons, elevons, elevators, ailerons, flaps, brakes or slats of an aerial vehicle, within desired ranges. In some embodiments, the control system 220 may be integrated with one or more of the processor 212, the memory 214 and/or the transceiver 216.
The motors 222 may be any type or form of motor (e.g., electric, gasoline-powered or any other type of motor) capable of generating sufficient rotational speeds of one or more propellers or other components to provide thrust and/or lift forces to the aerial vehicle 210 and any payload engaged thereby. In some embodiments, one or more of the motors 222 may be a brushless direct current (“DC”) multi-phase motor such as an outrunner brushless motor or an inrunner brushless motor.
The aerial vehicle 210 may include any number of such motors 222 of any kind. For example, one or more of the motors 222 may be aligned or configured to operate with different capacities or ratings, or at different speeds, or coupled to any number of propellers having different sizes and shapes. Additionally, one or more of the motors 222 may be an electric motor, e.g., a brushless DC multi-phase motor, and one or more of the motors 222 may be a gasoline-powered motor.
The aerial vehicle 210 may also include any number of sensors 230 or other components or other features for capturing data within a vicinity of the aerial vehicle 210, including but not limited to one or more imaging devices (e.g., digital cameras), range sensors (e.g., LIDAR sensors, time-of-flight sensors, ultrasonic sensors, or others), Global Positioning System (“GPS”) receivers or sensors, compasses, speedometers, altimeters, thermometers, barometers, hygrometers, gyroscopes, air monitoring sensors (e.g., oxygen, ozone, hydrogen, carbon monoxide or carbon dioxide sensors), ozone monitors, pH sensors, magnetic anomaly detectors, metal detectors, radiation sensors (e.g., Geiger counters, neutron detectors, alpha detectors), attitude sensors, depth gauges, accelerometers, or sound sensors (e.g., microphones, piezoelectric sensors, vibration sensors or other transducers for detecting and recording acoustic energy from one or more directions). For example, in some implementations, the aerial vehicle 210 may include two or more of such sensors 230 in an IMU or any other system.
As is also shown in
The conductive components 265 may have any size, shape or dimensions (e.g., lengths, widths or thicknesses), and may be formed from any suitable materials, including but not limited to copper, aluminum or any other conductive or semi-conductive materials. The sizes, shapes and dimensions of the conductive components 265, or the materials from which the conductive components 265 are formed, may be selected on any basis, e.g., to increase or decrease sensitivity of the capacitive sensors 260, which may be defined or determined based on the size, the shape, the dimension or the dielectric properties of the portions of the aerial vehicle 210 to which the conductive components 265 are joined, or any other factors.
Although the block diagram of
Furthermore, in some implementations, the motors 222, sensors 230, the capacitive sensors 260 and other components may be provided in association with systems other than aerial vehicles. For example, the motors 222, sensors 230, the capacitive sensors 260 and other components described herein may be utilized in connection with any systems having motor-driven or powered components that rotate or otherwise move.
The data processing system 280 includes one or more physical computer servers 282 having one or more computer processors, one or more data stores (e.g., databases) 284, and any number of transceivers 286 associated therewith, as well as provided for any specific or general purpose. In some embodiments, the data processing system 280 of
The servers 282 may be connected to or otherwise communicate with the data stores 284 and the transceivers 286, and may receive, analyze or store any type of information or data, e.g., imaging data, acoustic signals, environmental conditions, operational characteristics, or positions, or any other information or data, for any purpose. The servers 282 and/or the data stores 284 may also connect to or otherwise communicate with the network 290, through the sending and receiving of digital data. In some embodiments, the data processing system 280 may be provided in a physical location. In other such embodiments, the data processing system 280 may be provided in one or more alternate or virtual locations, e.g., in a “cloud”-based environment. In still other embodiments, the data processing system 280 may be provided onboard one or more vehicles, including but not limited to the aerial vehicle 210.
Each of the transceivers 216, 286 or other communications devices, systems or components may be configured to communicate through one or more wired or wireless means, e.g., wired technologies such as Universal Serial Bus (or “USB”) or fiber optic cable, or wireless standards or protocols such as Bluetooth or any Wi-Fi protocol, over the network 290 or directly. The transceivers 216, 286 may further include or be in communication with one or more I/O interfaces, network interfaces or devices, and may be configured to allow information or data to be exchanged between one another, or to or from one or more other computer devices or systems via the network 290.
The transceivers 216, 286 may perform any necessary protocol, timing or other data transformations in order to convert data signals from a first format suitable for use by one component into a second format suitable for use by another component. For example, in some embodiments, the transceivers 216, 286 may be configured to communicate according to one or more protocols traditionally associated with discrete components, e.g., within specific frequency spectra, including but not limited to radio frequency signals typically within ranges of approximately three hundred to four hundred megahertz, or MHz, as well as radio frequency signals typically within ranges of approximately three hundred to five hundred megahertz, or MHz.
In some embodiments, the transceivers 216, 286 may be configured to transmit or receive Bluetooth signals sent within an industrial, scientific and medical (ISM) radio frequency range of approximately 2.400 to approximately 2.500 gigahertz (GHz), with a center frequency of approximately 2.450 GHz. Such signals are typically sent within a particular range of approximately 2.402 to approximately 2.480 GHz, and have a maximum transmission rate on Bluetooth networks of 3 megabits per second (or 3 Mbps) or more, with a maximum transmission range of 10 to 100 meters or more, and at powers of approximately one milliwatt (mW), or 0.001 watts (W). The transceivers 216, 286 may operate according to a frequency-hopping spread spectrum (FHSS) method, e.g., by switching carriers across multiple frequency channels and transmitting communications in small segments, and in respective time slots. In some embodiments, such communications may be transmitted at approximately 1,600 times per second across 79 different subfrequencies at bandwidths of 0.001 GHz, or one megahertz (MHz) across the 2.402 to 2.480 GHz range.
In some embodiments, the transceivers 216, 286 may be configured to transmit or receive Ultra-Wide Band (or “UWB”) signals over one or more broad spectrums within any radio frequency range. In some embodiments, the UWB signals may be transmitted and/or received over a range of approximately three to eleven gigahertz (GHz), or over a range of approximately four to six gigahertz (GHz), with a center frequency of approximately five gigahertz (GHz). Alternatively, the UWB signals may be transmitted over any frequency range, and with any center frequency. Additionally, the UWB signals may consist of billions of pluses transmitted across such frequency spectrums, and at various power levels or power densities.
In some embodiments, the transceivers 216, 286 may be configured to transmit or receive RFID signals at low frequency (LF), medium frequency (MF) or high frequency (HF) levels (e.g., approximately 3 kilohertz to 30 megahertz, or 3 kHz-30 MHz) and transfer relatively small-sized sets or packets of data over short ranges (e.g., between one and one hundred centimeters, or 1-100 cm). In some other embodiments, the transceivers 216, 286 may be configured to transmit or receive RFID signals at higher frequency levels, such as ultrahigh frequency (UHF) or microwave levels (e.g., approximately 300 megahertz to 300 gigahertz, or 300 MHz-300 GHz) including larger sets or packets of data at ranges of one meter (1 m) or longer.
Alternatively, the transceivers 216, 286 may be configured to communicate within customized frequency spectra, or at dynamic frequencies, in accordance with the present disclosure. For example, in some embodiments, one or more of the transceivers 216, 286 may be configured to communicate according to a proprietary standard or protocol, or within a proprietary frequency spectrum.
In some embodiments, the transceivers 216, 286 may include support for devices attached through various types of peripheral buses, e.g., variants of the Peripheral Component Interconnect (PCI) bus standard or the Universal Serial Bus (USB) standard. In some other embodiments, functions of the transceivers 216, 286 may be split into two or more separate components, or incorporated directly into one or more processors or other components.
The network 290 may be any wired network, wireless network, or combination thereof, and may comprise the Internet in whole or in part. In addition, the network 290 may be a personal area network, local area network, wide area network, cable network, satellite network, cellular telephone network, or combination thereof. The network 290 may also be a publicly accessible network of linked networks, possibly operated by various distinct parties, such as the Internet. In some embodiments, the network 290 may be a private or semi-private network, such as a corporate or university intranet. The network 290 may include one or more wireless networks, such as a Global System for Mobile Communications (GSM) network, a Code Division Multiple Access (CDMA) network, a Long-Term Evolution (LTE) network, or some other type of wireless network. Protocols and components for communicating via the Internet or any of the other aforementioned types of communication networks are well known to those skilled in the art of computer communications and thus, need not be described in more detail herein. Any combination of networks or communications protocols may be utilized in accordance with the systems and methods of the present disclosure. For example, the various components described herein may be configured to communicate via an open or standard protocol such as Wi-Fi. Alternatively, such components may be configured to communicate with one another directly outside of a centralized network, such as the network 290, e.g., by a wireless protocol such as Bluetooth, in which two or more of such components may be paired with one another.
The computers, servers, devices and the like described herein have the necessary electronics, software, memory, storage, databases, firmware, logic/state machines, microprocessors, communication links, displays or other visual or audio user interfaces, printing devices, and any other I/O devices to provide any of the functions or services described herein and/or achieve the results described herein. Also, those of ordinary skill in the pertinent arts will recognize that users of such computers, servers, devices and the like may operate a keyboard, keypad, mouse, stylus, touch screen, or other device (not shown) or method to interact with the computers, servers, devices and the like, or to “select” an item, link, node, hub or any other aspect of the present disclosure.
The aerial vehicle 210 and/or the data processing system 280 may use any web-enabled or Internet applications or features, or any other client-server applications or features including electronic mail (or E-mail), or other messaging techniques, to connect to the network 290 or to communicate with one another, such as through short or multimedia messaging service (SMS or MMS) text messages, social network messages, online marketplace messages, telephone calls or the like. For example, the aerial vehicle 210 and/or the data processing system 280 may be adapted to transmit information or data in the form of synchronous or asynchronous messages between or among themselves, or between or among any other computer device in real time or in near-real time, or in one or more offline processes, via the network 290. Those of ordinary skill in the pertinent arts would recognize that the aerial vehicle 210 and/or the data processing system 280 may operate any of a number of computing devices that are capable of communicating over the network, including but not limited to set-top boxes, personal digital assistants, digital media players, web pads, laptop computers, desktop computers, laptop computers, tablet computers, smartphones, smart speakers, wrist watches, electronic book readers, and the like. The protocols and components for providing communication between such devices are well known to those skilled in the art of computer communications and need not be described in more detail herein.
The data and/or computer-executable instructions, programs, firmware, software and the like (also referred to herein as “computer-executable” components) described herein may be stored on a computer-readable medium that is within or accessible by computers or computer components such as the processors 212 and/or the data processing system 280, or any other computers or control systems, and having sequences of instructions which, when executed by a processor (e.g., a central processing unit, or “CPU”), cause the processor to perform all or a portion of the functions, services and/or methods described herein. Such computer-executable instructions, programs, software and the like may be loaded into the memory of one or more computers using a drive mechanism associated with the computer readable medium, such as a floppy drive, CD-ROM drive, DVD-ROM drive, network interface, or the like, or via external connections.
Some embodiments of the systems and methods of the present disclosure may also be provided as a computer-executable program product including a non-transitory machine-readable storage medium having stored thereon instructions (in compressed or uncompressed form) that may be used to program a computer (or other electronic device) to perform processes or methods described herein. The machine-readable storage medium may include, but is not limited to, hard drives, floppy diskettes, optical disks, CD-ROMs, DVDs, ROMs, RAMs, erasable programmable ROMs (“EPROM”), electrically erasable programmable ROMs (“EEPROM”), flash memory, magnetic or optical cards, solid-state memory devices, or other types of media/machine-readable medium that may be suitable for storing electronic instructions. Further, embodiments may also be provided as a computer-executable program product that includes a transitory machine-readable signal (in compressed or uncompressed form). Examples of machine-readable signals, whether modulated using a carrier or not, may include, but are not limited to, signals that a computer system or machine hosting or running a computer program can be configured to access, or including signals that may be downloaded through the Internet or other networks.
As is discussed above, systems and methods of the present disclosure may include capacitive sensing systems having circuits including one or more components for detecting levels of capacitance on conductive components provided on surfaces of aerial vehicles or other systems. Referring to
As is shown in
The conductive component 365 may be coupled to the integrated circuit component 362 in a manner that enables the capacitive sensor 360 to determine a level or change in a level of capacitance Cx on the conductive component 365. The integrated circuit component 362 may thus generate and transmit one or more signals representative of the level of capacitance Cx, or of any changes in the level of capacitance Cx, to a control system for operating one or more motors or other systems for causing movements of one or more objects. The conductive component 365 may be formed from any suitable materials, e.g., copper, aluminum or any other conductive or semi-conductive materials, and applied to surfaces of systems in any manner, such as by one or more fasteners, such as clips, screws, bolts, snap-fit connectors, or any other like features, or one or more adhesives, such as glues, tapes or other materials. Alternatively, the conductive component 365 may be a portion of such systems. For example, in some implementations, the conductive component 365 may be an integral portion of a ring, a spoke, or any other portion of a mesh section provided in association with a housing of an aerial vehicle or another system.
The integrated circuit component 362 may have one or more leads connected to a capacitor (e.g., a charge-sensing sample capacitor) 364, a series resistor 366 and a voltage source 368. Alternatively, or additionally, the capacitive sensor 360 may include any components other than the integrated circuit component 362 for determining a level or change of capacitance Cx on the conductive component 365. Moreover, in some implementations, the integrated circuit component 362 (or any other component) may receive inputs from any number of conductive components 365, and may be configured to determine or generate signals representative of levels or change of capacitance on any of such components. Alternatively, in some implementations, an aerial vehicle or another system may include any number of integrated circuit components 362 (or other components), each configured to receive inputs from one or more conductive components 365, and generate signals representative of levels or change of capacitance on any of such components.
In accordance with one or more implementations, the integrated circuit component 362 may include any number of transistors, e.g., small-outline transistors, or transistors of any other type or form.
The capacitor 364 may have any value or level of capacitance CS, e.g., on an order of one to fifty nanofarads (1 to 50 nF), or one to one hundred nanofarads (1 to 100 nF), that may be selected to configure or enhance a level of sensitivity of the capacitive sensor 360 or to vary a level of response time of the capacitive sensor 360. For example, the capacitor 364 may have a value or level of capacitance CS selected based on attributes of surfaces of a system with which the conductive component 365 is associated, such as thicknesses or dielectric properties of such surfaces. Where a value or level of capacitance CS of the capacitor 364 is substantially high, greater stability and higher dielectric properties of such surfaces are required in order to enhance the sensitivity and reliability of signals generated by the integrated circuit component 362. Preferably, a value or level of capacitance CS is selected to be approximately one thousand times greater than an anticipated range in the value or level of capacitance Cx.
The capacitor 364 may be of any type, and may take any shape or form. In some implementations, the capacitor 364 may be a stable-type capacitor, such as an X7R capacitor, a polyphenylene sulfide film capacitor, or any other type or form of capacitor.
The series resistor 366 may be provided in line with a connection between the conductive component 365 and the integrated circuit component 362, in order to limit any electrostatic discharge or suppress any interference, e.g., radio frequency interference. The series resistor 366 may have any value or level of resistance RS, e.g., approximately one to forty kiloohms (1 to 40 kΩ). Alternatively, in some implementations, the series resistor 366 may be omitted from the capacitive sensor 360.
The voltage source 368 may be any source of direct current electrical power for the capacitive sensor 360. For example, the voltage source 368 may be an independent power supply for the capacitive sensor 360, or a connection to a power supply that is shared by the capacitive sensor 360 and any other components of a system, such as an aerial vehicle. In some implementations, the voltage source 368 may include any number of regulators, switches, diodes or other systems for supplying, controlling or isolating power to the capacitive sensor 360.
In addition to the integrated circuit component 362, the capacitor 364, the resistor 366 and the voltage source 368, the capacitive sensor 360 may include any number of other components (not shown in
Referring to
A control system or another component of the aerial vehicle may be programmed with one or more sets of instructions for generating trajectories to cause the aerial vehicle to travel on such courses, speeds or altitudes along the route, or to perform such actions. For example, the aerial vehicle may be configured to calculate trajectories for traveling along a route or path defined by waypoints, which may be represented in global or local coordinates, e.g., Cartesian coordinates, such as in a manner that minimizes snap. An aerial vehicle may calculate a trajectory at any time prior to, during or following operations, such as when the aerial vehicle is at or over a waypoint, when the aerial vehicle has potentially detected an obstacle, or when the aerial vehicle has determined that a braking maneuver is necessary, among others. A trajectory calculated in this manner may include sets of points in three-dimensional space, e.g., a set of points in a three-dimensional coordinate system.
In some implementations, a trajectory may also include, for each defined point in three-dimensional space forming part of the trajectory, a time value indicating a time at which the trajectory passes through that point, or a time at which an aerial vehicle following that trajectory should pass through that point. Thus, a trajectory may be characterized as defining points in four dimensions, viz., in three-dimensional space and time. In some implementations, a trajectory may also include, for each defined point in three-dimensional space forming part of the trajectory, a yaw value indicating a yaw angle of an aerial vehicle at that point of the trajectory. In some implementations, a trajectory is calculated using, or is defined by, one or more equations, e.g., a set of polynomial equations, and positions of points along the trajectory may be calculated using the one or more trajectory equations. In some implementations, one or more trajectory equations may be used to generate x(t), y(t), and z(t) values indicating a position of the aerial vehicle in three-dimensional space at each time t along a trajectory. In some implementations, the trajectory equations, or the x(t), y(t), and z(t) values generated based thereon, may also be used to generate velocity information for the aerial vehicle in three-dimensional space at each time t along the trajectory, including both a speed and a direction of the aerial vehicle at each time t.
In some implementations, one or more trajectory equations may be used to generate a yaw value ψ(t) indicating a yaw angle of the aerial vehicle at the time t along a trajectory. In some implementations, trajectory equations or yaw values generated based thereon may be used to generate yaw angular velocity information for the aerial vehicle at each time t along the trajectory, including both a rotational speed and a rotational direction of the aerial vehicle at each time t.
The capacitive sensor may include any number of capacitors, as well as any number of components for detecting levels of capacitance, or changes in such levels, of one or more capacitors of a capacitive sensor. For example, in some implementations, the aerial vehicle may include a safety mesh or grill covering portions of propeller, motors or other components that may rotate or otherwise move during operations of the aerial vehicle. Alternatively, the safety mesh or grill may cover components that reach extreme temperatures (e.g., very hot or very cold surfaces), sharp components, or components that should not be contacted by humans or other animals during such operations.
The capacitive sensor may, therefore, include one or more components for sensing the proximity of portions of such humans or other animals to components of the aerial vehicle during operations. In some implementations, the capacitive sensor may include an electrode provided in association with a portion of the mesh or grill, which may be conductive or partially conductive in nature, and formed from one or more conductive or semi-conductive materials such as aluminum, carbon fiber, copper, graphite, or others. In some implementations, the capacitive sensor may further include one or more regulators or integrated circuits or integrated circuit components. For example, in some implementations, the capacitive sensor may include a conductive portion of a mesh, a grill or another protective feature, as well as an integrated circuit component coupled to a regulator, and a ground plane arranged beneath the regulator and a ground pin of the integrated circuit component.
The capacitive sensor may also include one or more components or features having any number of attributes that may be selected to maintain a level of sensitivity of the capacitive sensor. For example, the capacitive sensor may include one or more reference capacitors and/or resistors having values of capacitance or resistance that are selected to maintain a level of gain of the capacitive sensor or a response time of the capacitive sensor. Alternatively, the values of capacitance or resistance may be selected based on dielectric constants of materials of the aerial vehicle to which the electrodes are affixed. Alternatively, the capacitances or resistances on the reference capacitors and resistors may be selected on any other basis.
Alternatively, the aerial vehicle may include any number of capacitive sensors, and need not be limited to a single capacitive sensor.
Furthermore, as yet another alternative, the capacitive sensors may be provided on a system other than an aerial vehicle within a vicinity of one or more objects or components that rotate or otherwise move during operations. For example, the capacitive sensors may be mounted to a blender, a food processor, a lawn mower or trimmer, or any other such system.
At box 420, one or more capacitors of the capacitive sensor coupled to the conductive components are charged during the operations of the aerial vehicle. For example, where the capacitive sensor includes a voltage supply (e.g., a DC voltage supply), voltage may be applied to the one or more capacitors while the aerial vehicle is in flight or, alternatively, prior to taking off from an origin, such as during pre-flight preparations. The capacitive sensor may include any number of switches or other controllers that may be operated to apply voltage to the one or more capacitors, or to isolate voltage from the one or more capacitors.
At box 430, a level of capacitance on the conductive components of the capacitive sensor is determined. The level of capacitance may be determined in any manner. For example, in some implementations, the capacitors may be constantly discharged and charged during the operations of the aerial vehicle, and a time required to discharge the capacitors may be determined. A level of capacitance on the conductive components may be determined as a function of a discharge time or a charge time.
At box 440, whether a change in the level of capacitance in excess of a predetermined threshold has been detected is determined. For example, the level of capacitance on the one or more conductive components may be determined continuously, or at any predetermined interval of time. Where an overall change in the levels of capacitance exceeds a predetermined threshold, or whether a rate of change in the levels of capacitance exceeds the predetermined threshold, may be determined based on consecutive levels of capacitance determined at box 430, or levels of capacitance determined at various times. For example, where times required to discharge or charge the capacitors during operations are determined, a change in a level of capacitance may be detected where a change in either a time required to discharge the capacitors or a time required to charge the capacitors is observed, and the change may be compared to one or more thresholds of time or capacitance that are selected in any manner and on any basis. For example, a predetermined threshold may be selected based on a level of sensitivity of the capacitive sensor, a desired response time for the capacitive sensor, dimensions of one or more indoor spaces in which the aerial vehicle is operating, or any other factors.
If a change in the level of capacitance exceeding the predetermined threshold has not been detected at box 440, then the process advances to box 450, where whether the aerial vehicle has arrived at a destination is determined. For example, where the aerial vehicle is configured with one or more sensors for determining positions within indoor spaces, e.g., position sensors, cameras, odometers or others, the aerial vehicle may capture data using such sensors and compare a position determined based on such data to a position associated with the destination to determine whether the aerial vehicle is sufficiently proximate the destination in order to engage in one or more landing operations, or has slowed, hovered, braked or landed at the destination.
If the aerial vehicle has not arrived at the destination, then the process returns to box 420, where one or more capacitors of the capacitive sensor are charged during the continued operations of the aerial vehicle. If the aerial vehicle has arrived at the destination, however, then the process advances to box 460, where the aerial vehicle lands at the destination, and the process ends.
If a change in the level of capacitance in excess of the predetermined threshold has been detected at box 440, however, then the process advances to box 470, where a detection of a body part within a proximity of the aerial vehicle is recorded. For example, where the change in the level of capacitance of the one or more capacitors is determined to have exceeded a predetermined threshold, the change may be presumed to be associated with a presence of a body part of a human or another animal within a vicinity of one of the electrodes.
Upon detecting the body part, information or data regarding the operations of the aerial vehicle may be determined and stored in one or more data stores. For example, in some implementations, a position of the aerial vehicle at a time of the detection, as well as a course, a speed and an altitude on which the aerial vehicle was traveling at the time, may be determined and recorded. Alternatively, or additionally, any additional information or data, such as images or other data captured at or near the time of the detection may be stored in association with the position, the course, the speed or the altitude. Such information or data may also identify one or more operations being performed by the aerial vehicle at the time of the detection.
At box 480, the aerial vehicle executes one or more actions in response to the detection of the body part, and the process ends. For example, the aerial vehicle may be programmed or configured to, upon detecting a body part within a vicinity based on a change in a level of capacitance, execute one or more actions in response to the detection. Such actions may include, but need not be limited to, causing the aerial vehicle to travel from a position at which the change was detected to a position that is located at a safe distance, range or altitude above, away or from the body part. For example, the aerial vehicle may be preprogrammed with one or more sets of instructions that, when executed, cause the aerial vehicle to operate one or more of motors to generate lift and thrust to cause the aerial vehicle to change its position by predetermined amounts (or distances) along one or more principal axes. Values of the amounts (or distances) may be selected on any basis, such as values of the changes or one or more safety factors, as well as any dimensions or features of the indoor spaces. Alternatively, in some implementations, the aerial vehicle may execute one or more sets of instructions to cause one or more of the motors to immediately stop, thereby causing the aerial vehicle to descend in altitude, away from the position at which the change was detected. While falling, the aerial vehicle may sense the change in altitude and execute one or more sets of instructions to cause the motors to reengage in operation once the aerial vehicle is sufficiently far from the position at which the change was detected, and to land or otherwise take any other action. As yet another alternative, an aerial vehicle may generate one or more visual or audible cues, such as by activating any number of illuminators to project light in any number of colors therefrom, or by causing one or more sounds (e.g., tones, sirens or spoken words such as “stay away!”) to be emitted from one or more speakers, or any other actions.
Thresholds for initiating changes in operations of systems upon detecting changes in levels of capacitance may be selected on any basis. Referring to
As is shown in
The conductive components 565 and any capacitors, resistors or other components of the capacitive sensing system 560 may be selected on any basis in order to achieve a level of sensitivity that enables the aerial vehicle 510 to respond to a detection of the presence of one or more body parts of humans or other animals, e.g., a hand 575 of a grounded human body, as necessary. For example, sizes, shapes or dimensions of the conductive components 565, as well as materials from which the conductive components 565 are formed or locations on the aerial vehicle 510 on which the conductive components 565 are mounted or affixed, may be selected to ensure that the capacitive sensing system 560 detects the presence of one or more body parts at a predetermined distance h from the aerial vehicle 510 in general, or from one or more of the conductive components 565 in particular. Alternatively, or additionally, capacitors, resistors, transistors or other components of the capacitive sensing system 560 may be selected to establish a level of gain or sensitivity of the capacitive sensing system 560, or on any other basis.
As is further shown in
As is shown in
Similarly, as is shown in
Finally, as is shown in
For example, upon detecting that a change ΔC in capacitance has been detected, the aerial vehicle 510 automatically initiates a change its position by descending by at least a predetermined distance ΔP, such as by isolating power to motors of the aerial vehicle 510, or by instructing such motors to operate in a manner that causes the aerial vehicle 510 to descend by the distance ΔP. In some implementations, the aerial vehicle 510 may operate such motors to cause a change in position that is vertically above or laterally departed from a position at which the maximum change ΔC in capacitance was detected. Alternatively, the aerial vehicle 510 may take any other action upon determining that the change ΔC in capacitance was detected.
The systems and methods of the present disclosure are not limited for use in connection with aerial vehicles. Referring to
As is shown in
As is shown in
Referring to
As is shown in
As is shown in
The contents of International Patent Publication No. WO 2021/061810 are incorporated by reference herein in their entirety.
Although some embodiments of the present disclosure show the use of autonomous vehicles in support of one or more applications or functions (e.g., security applications or functions) within indoor spaces of a facility such as a home or a like structure, those of ordinary skill in the pertinent arts will realize that the systems and methods of the present disclosure are not so limited. Rather, the systems and methods of the present disclosure may be utilized in connection with any facility, including but not limited to homes, in support of any type of application or function.
It should be understood that, unless otherwise explicitly or implicitly indicated herein, any of the features, characteristics, alternatives or modifications described regarding a particular embodiment herein may also be applied, used, or incorporated with any other embodiment described herein, and that the drawings and detailed description of the present disclosure are intended to cover all modifications, equivalents and alternatives to the various embodiments as defined by the appended claims. Moreover, with respect to the one or more methods or processes of the present disclosure described herein, such as the flow chart of
Conditional language, such as, among others, “can,” “could,” “might,” or “may,” unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey in a permissive manner that certain embodiments could include, or have the potential to include, but do not mandate or require, certain features, elements and/or steps. In a similar manner, terms such as “include,” “including” and “includes” are generally intended to mean “including, but not limited to.” Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment.
The elements of a method, process, or algorithm described in connection with the embodiments disclosed herein can be embodied directly in hardware, in a software module stored in one or more memory devices and executed by one or more processors, or in a combination of the two. A software module can reside in RAM, flash memory, ROM, EPROM, EEPROM, registers, a hard disk, a removable disk, a CD-ROM, a DVD-ROM or any other form of non-transitory computer-readable storage medium, media, or physical computer storage known in the art. An example storage medium can be coupled to the processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium can be integral to the processor. The storage medium can be volatile or nonvolatile. The processor and the storage medium can reside in an ASIC. The ASIC can reside in a user terminal. In the alternative, the processor and the storage medium can reside as discrete components in a user terminal.
Disjunctive language such as the phrase “at least one of X, Y, or Z,” or “at least one of X, Y and Z,” unless specifically stated otherwise, is otherwise understood with the context as used in general to present that an item, term, etc., may be either X, Y, or Z, or any combination thereof (e.g., X, Y, and/or Z). Thus, such disjunctive language is not generally intended to, and should not, imply that certain embodiments require at least one of X, at least one of Y, or at least one of Z to each be present.
Unless otherwise explicitly stated, articles such as “a” or “an” should generally be interpreted to include one or more described items. Accordingly, phrases such as “a device configured to” are intended to include one or more recited devices. Such one or more recited devices can also be collectively configured to carry out the stated recitations. For example, “a processor configured to carry out recitations A, B and C” can include a first processor configured to carry out recitation A working in conjunction with a second processor configured to carry out recitations B and C.
Language of degree used herein, such as the terms “about,” “approximately,” “generally,” “nearly” or “substantially” as used herein, represent a value, amount, or characteristic close to the stated value, amount, or characteristic that still performs a desired function or achieves a desired result. For example, the terms “about,” “approximately,” “generally,” “nearly” or “substantially” may refer to an amount that is within less than 10% of, within less than 5% of, within less than 1% of, within less than 0.1% of, and within less than 0.01% of the stated amount.
Although the invention has been described and illustrated with respect to illustrative embodiments thereof, the foregoing and various other additions and omissions may be made therein and thereto without departing from the spirit and scope of the present disclosure.
Number | Name | Date | Kind |
---|---|---|---|
3417729 | Gilday et al. | Dec 1968 | A |
3575527 | Sumitani et al. | Apr 1971 | A |
3805723 | Bernaerts | Apr 1974 | A |
4865248 | Barth | Sep 1989 | A |
4954962 | Evans, Jr. et al. | Sep 1990 | A |
5040116 | Evans, Jr. et al. | Aug 1991 | A |
5283739 | Summerville et al. | Feb 1994 | A |
5371581 | Wangler et al. | Dec 1994 | A |
5386462 | Schlamp | Jan 1995 | A |
5452374 | Cullen et al. | Sep 1995 | A |
5497236 | Wolff et al. | Mar 1996 | A |
5731884 | Inoue | Mar 1998 | A |
5847522 | Barba | Dec 1998 | A |
5901253 | Tretter | May 1999 | A |
5995898 | Tuttle | Nov 1999 | A |
6031612 | Shirley | Feb 2000 | A |
6266577 | Popp et al. | Jul 2001 | B1 |
6344796 | Ogilvie et al. | Feb 2002 | B1 |
6374155 | Wallach et al. | Apr 2002 | B1 |
6426699 | Porter | Jul 2002 | B1 |
6507670 | Moed | Jan 2003 | B1 |
6543983 | Felder et al. | Apr 2003 | B1 |
6636781 | Shen et al. | Oct 2003 | B1 |
6676460 | Motsenbocker | Jan 2004 | B1 |
6690997 | Rivalto | Feb 2004 | B2 |
6694217 | Bloom | Feb 2004 | B2 |
6705523 | Stamm et al. | Mar 2004 | B1 |
6804607 | Wood | Oct 2004 | B1 |
6919803 | Breed | Jul 2005 | B2 |
6954290 | Braudaway et al. | Oct 2005 | B1 |
6961711 | Chee | Nov 2005 | B1 |
6965440 | Nakagiri et al. | Nov 2005 | B1 |
6970838 | Kamath et al. | Nov 2005 | B1 |
7006952 | Matsumoto et al. | Feb 2006 | B1 |
7016536 | Ling et al. | Mar 2006 | B1 |
7031519 | Elmenhurst | Apr 2006 | B2 |
7129817 | Yamagishi | Oct 2006 | B2 |
7133743 | Tilles et al. | Nov 2006 | B2 |
7145699 | Dolan | Dec 2006 | B2 |
7149611 | Beck et al. | Dec 2006 | B2 |
7171879 | Gass et al. | Feb 2007 | B2 |
7188513 | Wilson | Mar 2007 | B2 |
7335071 | Motsenbocker | Feb 2008 | B1 |
7337686 | Sagi-Dolev | Mar 2008 | B2 |
7337944 | Devar | Mar 2008 | B2 |
7339993 | Brooks et al. | Mar 2008 | B1 |
7459880 | Rosen | Dec 2008 | B1 |
7639386 | Siegel et al. | Dec 2009 | B1 |
7668404 | Adams et al. | Feb 2010 | B2 |
7673831 | Steele et al. | Mar 2010 | B2 |
7685953 | Giles | Mar 2010 | B2 |
7693745 | Pomerantz et al. | Apr 2010 | B1 |
7894939 | Zini et al. | Feb 2011 | B2 |
7925375 | Schininger et al. | Apr 2011 | B2 |
7946526 | Zimet | May 2011 | B2 |
7946530 | Talmage, Jr. | May 2011 | B1 |
7966093 | Zhuk | Jun 2011 | B2 |
8015023 | Lee et al. | Sep 2011 | B1 |
8078317 | Allinson et al. | Dec 2011 | B2 |
8126642 | Trepagnier et al. | Feb 2012 | B2 |
8131607 | Park et al. | Mar 2012 | B2 |
8145351 | Schininger et al. | Mar 2012 | B2 |
8195328 | Mallett et al. | Jun 2012 | B2 |
8286236 | Jung et al. | Oct 2012 | B2 |
8412588 | Bodell et al. | Apr 2013 | B1 |
8418959 | Kang et al. | Apr 2013 | B2 |
8429754 | Jung et al. | Apr 2013 | B2 |
8473189 | Christoph | Jun 2013 | B2 |
8511606 | Lutke et al. | Aug 2013 | B1 |
8577538 | Lenser et al. | Nov 2013 | B2 |
8599027 | Sanchez | Dec 2013 | B2 |
8602349 | Petrov | Dec 2013 | B2 |
8639400 | Wong | Jan 2014 | B1 |
8736820 | Choe et al. | May 2014 | B2 |
8752166 | Jung et al. | Jun 2014 | B2 |
8791790 | Robertson et al. | Jul 2014 | B2 |
8874301 | Rao et al. | Oct 2014 | B1 |
8899903 | Saad et al. | Dec 2014 | B1 |
8948914 | Zini et al. | Feb 2015 | B2 |
8956100 | Davi et al. | Feb 2015 | B2 |
8989053 | Skaaksrud et al. | Mar 2015 | B1 |
9033285 | Iden et al. | May 2015 | B2 |
9051043 | Peeters et al. | Jun 2015 | B1 |
9079587 | Rupp et al. | Jul 2015 | B1 |
9139310 | Wang | Sep 2015 | B1 |
9163909 | Chengalva | Oct 2015 | B2 |
9193452 | Carreker | Nov 2015 | B2 |
9195959 | Lopez et al. | Nov 2015 | B1 |
9216587 | Ando et al. | Dec 2015 | B2 |
9216857 | Kalyan et al. | Dec 2015 | B1 |
9230236 | Villamar | Jan 2016 | B2 |
9235213 | Villamar | Jan 2016 | B2 |
9244147 | Soundararajan et al. | Jan 2016 | B1 |
9256852 | Myllymaki | Feb 2016 | B1 |
9261578 | Im et al. | Feb 2016 | B2 |
9321531 | Takayama et al. | Apr 2016 | B1 |
9336506 | Shucker et al. | May 2016 | B2 |
9336635 | Robertson et al. | May 2016 | B2 |
9358975 | Watts | Jun 2016 | B1 |
9373149 | Abhyanker | Jun 2016 | B2 |
9381916 | Zhu et al. | Jul 2016 | B1 |
9397518 | Theobald | Jul 2016 | B1 |
9404761 | Meuleau | Aug 2016 | B2 |
9409644 | Stanek et al. | Aug 2016 | B2 |
9411337 | Theobald et al. | Aug 2016 | B1 |
9412280 | Zwillinger et al. | Aug 2016 | B1 |
9436183 | Thakur et al. | Sep 2016 | B2 |
9436926 | Cousins et al. | Sep 2016 | B2 |
9448559 | Kojo et al. | Sep 2016 | B2 |
9489490 | Theobald | Nov 2016 | B1 |
9510316 | Skaaksrud | Nov 2016 | B2 |
9535421 | Canoso et al. | Jan 2017 | B1 |
9545852 | Streett | Jan 2017 | B2 |
9561941 | Watts | Feb 2017 | B1 |
9568335 | Thakur et al. | Feb 2017 | B2 |
9582950 | Shimizu et al. | Feb 2017 | B2 |
9600645 | Fadell et al. | Mar 2017 | B2 |
9619776 | Ford et al. | Apr 2017 | B1 |
9623553 | Theobald et al. | Apr 2017 | B1 |
9623562 | Watts | Apr 2017 | B1 |
9650136 | Haskin et al. | May 2017 | B1 |
9652912 | Fadell et al. | May 2017 | B2 |
9656805 | Evans et al. | May 2017 | B1 |
9671791 | Paczan | Jun 2017 | B1 |
9677564 | Woodworth et al. | Jun 2017 | B1 |
9682481 | Lutz et al. | Jun 2017 | B2 |
9697730 | Thakur et al. | Jul 2017 | B2 |
9718564 | Beckman et al. | Aug 2017 | B1 |
9720414 | Theobald | Aug 2017 | B1 |
9731821 | Hoareau et al. | Aug 2017 | B2 |
9733646 | Nusser et al. | Aug 2017 | B1 |
9746852 | Watts et al. | Aug 2017 | B1 |
9746853 | Scheepjens et al. | Aug 2017 | B2 |
9778653 | McClintock et al. | Oct 2017 | B1 |
9786187 | Bar-Zeev et al. | Oct 2017 | B1 |
9796529 | Hoareau et al. | Oct 2017 | B1 |
9828092 | Navot et al. | Nov 2017 | B1 |
9858604 | Apsley et al. | Jan 2018 | B2 |
9886035 | Watts et al. | Feb 2018 | B1 |
9896204 | Willison | Feb 2018 | B1 |
9959771 | Carlson | May 2018 | B1 |
9959773 | Raptopoulos et al. | May 2018 | B2 |
9974612 | Pinter et al. | May 2018 | B2 |
10022753 | Chelian et al. | Jul 2018 | B2 |
10022867 | Saboo et al. | Jul 2018 | B2 |
10048697 | Theobald | Aug 2018 | B1 |
10108185 | Theobald | Oct 2018 | B1 |
10137984 | Flick | Nov 2018 | B1 |
10558226 | Bigdeli | Feb 2020 | B1 |
10780988 | Buchmueller et al. | Sep 2020 | B2 |
11161608 | Namgoong | Nov 2021 | B2 |
11851162 | Daube | Dec 2023 | B1 |
12089385 | Zaghloul | Sep 2024 | B2 |
20010045449 | Shannon | Nov 2001 | A1 |
20020016726 | Ross | Feb 2002 | A1 |
20020035450 | Thackston | Mar 2002 | A1 |
20020072979 | Sinha et al. | Jun 2002 | A1 |
20020087375 | Griffin et al. | Jul 2002 | A1 |
20020107751 | Rajagopalan et al. | Aug 2002 | A1 |
20020111914 | Terada et al. | Aug 2002 | A1 |
20020116289 | Yang | Aug 2002 | A1 |
20020123930 | Boyd et al. | Sep 2002 | A1 |
20020156645 | Hansen | Oct 2002 | A1 |
20030040980 | Nakajima et al. | Feb 2003 | A1 |
20030072031 | Kuwata et al. | Apr 2003 | A1 |
20030121968 | Miller et al. | Jul 2003 | A1 |
20030141411 | Pandya et al. | Jul 2003 | A1 |
20040002898 | Kuhlmann et al. | Jan 2004 | A1 |
20040068416 | Solomon | Apr 2004 | A1 |
20040112660 | Johansson et al. | Jun 2004 | A1 |
20040160335 | Reitmeier et al. | Aug 2004 | A1 |
20040162638 | Solomon | Aug 2004 | A1 |
20040257199 | Fitzgibbon et al. | Dec 2004 | A1 |
20050068178 | Lee et al. | Mar 2005 | A1 |
20050093865 | Jia | May 2005 | A1 |
20050102240 | Misra et al. | May 2005 | A1 |
20050244060 | Nagarajan et al. | Nov 2005 | A1 |
20050285934 | Carter | Dec 2005 | A1 |
20060053534 | Mullen | Mar 2006 | A1 |
20060118162 | Saelzer et al. | Jun 2006 | A1 |
20060136237 | Spiegel et al. | Jun 2006 | A1 |
20060287829 | Pashko-Paschenko | Dec 2006 | A1 |
20070016496 | Bar et al. | Jan 2007 | A1 |
20070073552 | Hileman | Mar 2007 | A1 |
20070102565 | Speer et al. | May 2007 | A1 |
20070150375 | Yang | Jun 2007 | A1 |
20070170237 | Neff | Jul 2007 | A1 |
20070210953 | Abraham et al. | Sep 2007 | A1 |
20070233337 | Plishner | Oct 2007 | A1 |
20070244763 | Williams et al. | Oct 2007 | A1 |
20070246601 | Layton | Oct 2007 | A1 |
20070262195 | Bulaga et al. | Nov 2007 | A1 |
20070293978 | Wurman et al. | Dec 2007 | A1 |
20080012697 | Smith et al. | Jan 2008 | A1 |
20080027591 | Lenser et al. | Jan 2008 | A1 |
20080100258 | Ward | May 2008 | A1 |
20080109246 | Russell | May 2008 | A1 |
20080111816 | Abraham et al. | May 2008 | A1 |
20080141921 | Hinderks | Jun 2008 | A1 |
20080150679 | Bloomfield | Jun 2008 | A1 |
20080154659 | Bettes et al. | Jun 2008 | A1 |
20080167817 | Hessler et al. | Jul 2008 | A1 |
20080184906 | Kejha | Aug 2008 | A1 |
20080189012 | Kaufmann | Aug 2008 | A1 |
20080301009 | Plaster et al. | Dec 2008 | A1 |
20090027253 | Tooren et al. | Jan 2009 | A1 |
20090062974 | Tamamoto et al. | Mar 2009 | A1 |
20090063166 | Palmer | Mar 2009 | A1 |
20090079388 | Reddy | Mar 2009 | A1 |
20090086275 | Liang et al. | Apr 2009 | A1 |
20090091435 | Bolourchi | Apr 2009 | A1 |
20090106124 | Yang | Apr 2009 | A1 |
20090149985 | Chirnomas | Jun 2009 | A1 |
20090164379 | Jung et al. | Jun 2009 | A1 |
20090165127 | Jung et al. | Jun 2009 | A1 |
20090236470 | Goossen et al. | Sep 2009 | A1 |
20090254457 | Folsom | Oct 2009 | A1 |
20090254482 | Vadlamani et al. | Oct 2009 | A1 |
20090299903 | Hung et al. | Dec 2009 | A1 |
20090303507 | Abeloe | Dec 2009 | A1 |
20090314883 | Arlton et al. | Dec 2009 | A1 |
20100007479 | Smith | Jan 2010 | A1 |
20100030608 | Kaminsky et al. | Feb 2010 | A1 |
20100031351 | Jung et al. | Feb 2010 | A1 |
20100088163 | Davidson et al. | Apr 2010 | A1 |
20100088175 | Lundquist | Apr 2010 | A1 |
20100100269 | Ekhaguere et al. | Apr 2010 | A1 |
20100169185 | Cottingham | Jul 2010 | A1 |
20100206145 | Tetelbaum et al. | Aug 2010 | A1 |
20100287065 | Alivandi | Nov 2010 | A1 |
20100299067 | McCollough et al. | Nov 2010 | A1 |
20100299222 | Hamilton et al. | Nov 2010 | A1 |
20110035149 | McAndrew et al. | Feb 2011 | A1 |
20110074570 | Feldstein et al. | Mar 2011 | A1 |
20110087350 | Fogel et al. | Apr 2011 | A1 |
20110112761 | Hurley et al. | May 2011 | A1 |
20110153052 | Pettibone et al. | Jun 2011 | A1 |
20110166707 | Romanov et al. | Jul 2011 | A1 |
20110178711 | Christoph | Jul 2011 | A1 |
20110210866 | David et al. | Sep 2011 | A1 |
20110227435 | Maeda | Sep 2011 | A1 |
20110246331 | Luther et al. | Oct 2011 | A1 |
20110253831 | Cheng | Oct 2011 | A1 |
20110264311 | Lee et al. | Oct 2011 | A1 |
20110282476 | Hegemier et al. | Nov 2011 | A1 |
20110301787 | Chaperon et al. | Dec 2011 | A1 |
20110313878 | Norman | Dec 2011 | A1 |
20120039694 | Suzanne | Feb 2012 | A1 |
20120078592 | Sims | Mar 2012 | A1 |
20120091260 | Callou | Apr 2012 | A1 |
20120109419 | Mercado | May 2012 | A1 |
20120219397 | Baker | Aug 2012 | A1 |
20120221438 | Cook et al. | Aug 2012 | A1 |
20120227389 | Hinderks | Sep 2012 | A1 |
20120229325 | Dutruc | Sep 2012 | A1 |
20120234969 | Savoye et al. | Sep 2012 | A1 |
20120235606 | Takeuchi | Sep 2012 | A1 |
20120323365 | Taylor et al. | Dec 2012 | A1 |
20130006739 | Horvitz et al. | Jan 2013 | A1 |
20130073477 | Grinberg | Mar 2013 | A1 |
20130081245 | Vavrina et al. | Apr 2013 | A1 |
20130093582 | Walsh et al. | Apr 2013 | A1 |
20130126611 | Kangas et al. | May 2013 | A1 |
20130148123 | Hayashi | Jun 2013 | A1 |
20130206915 | Desaulniers | Aug 2013 | A1 |
20130218446 | Bradley et al. | Aug 2013 | A1 |
20130218799 | Lehmann et al. | Aug 2013 | A1 |
20130261792 | Gupta et al. | Oct 2013 | A1 |
20130262251 | Wan et al. | Oct 2013 | A1 |
20130262252 | Lakshman et al. | Oct 2013 | A1 |
20130262276 | Wan et al. | Oct 2013 | A1 |
20130262336 | Wan et al. | Oct 2013 | A1 |
20130264381 | Kim et al. | Oct 2013 | A1 |
20130320133 | Ratti et al. | Dec 2013 | A1 |
20130324164 | Vulcano | Dec 2013 | A1 |
20130332062 | Kreitmair-Steck et al. | Dec 2013 | A1 |
20140010656 | Nies | Jan 2014 | A1 |
20140022055 | Levien et al. | Jan 2014 | A1 |
20140025230 | Levien et al. | Jan 2014 | A1 |
20140030444 | Swaminathan et al. | Jan 2014 | A1 |
20140031964 | Sidhu et al. | Jan 2014 | A1 |
20140032034 | Raptopoulos et al. | Jan 2014 | A1 |
20140040065 | DuBois | Feb 2014 | A1 |
20140052661 | Shakes et al. | Feb 2014 | A1 |
20140058959 | Isbjornssund et al. | Feb 2014 | A1 |
20140081445 | Villamar | Mar 2014 | A1 |
20140089073 | Jacobs et al. | Mar 2014 | A1 |
20140136282 | Fedele | May 2014 | A1 |
20140136414 | Abhyanker | May 2014 | A1 |
20140149244 | Abhyanker | May 2014 | A1 |
20140156053 | Mahdavi et al. | Jun 2014 | A1 |
20140180914 | Abhyanker | Jun 2014 | A1 |
20140200697 | Cheng | Jul 2014 | A1 |
20140214684 | Pell | Jul 2014 | A1 |
20140244433 | Cruz | Aug 2014 | A1 |
20140254896 | Zhou et al. | Sep 2014 | A1 |
20140271200 | Sutton et al. | Sep 2014 | A1 |
20140283104 | Nilsson | Sep 2014 | A1 |
20140309813 | Ricci | Oct 2014 | A1 |
20140325218 | Shimizu et al. | Oct 2014 | A1 |
20140330456 | Morales et al. | Nov 2014 | A1 |
20150006005 | Yu et al. | Jan 2015 | A1 |
20150066178 | Stava | Mar 2015 | A1 |
20150069968 | Pounds | Mar 2015 | A1 |
20150098819 | Tourin et al. | Apr 2015 | A1 |
20150102154 | Duncan et al. | Apr 2015 | A1 |
20150112837 | O'Dea | Apr 2015 | A1 |
20150112885 | Fadell et al. | Apr 2015 | A1 |
20150120094 | Kimchi et al. | Apr 2015 | A1 |
20150120602 | Huffman et al. | Apr 2015 | A1 |
20150127712 | Fadell et al. | May 2015 | A1 |
20150129716 | Yoffe | May 2015 | A1 |
20150153175 | Skaaksrud | Jun 2015 | A1 |
20150154545 | Skaaksrud et al. | Jun 2015 | A1 |
20150158599 | Sisko | Jun 2015 | A1 |
20150175276 | Koster | Jun 2015 | A1 |
20150183528 | Walsh et al. | Jul 2015 | A1 |
20150185034 | Abhyanker | Jul 2015 | A1 |
20150191255 | Zolich et al. | Jul 2015 | A1 |
20150202770 | Patron et al. | Jul 2015 | A1 |
20150227882 | Bhatt | Aug 2015 | A1 |
20150246727 | Masticola et al. | Sep 2015 | A1 |
20150253777 | Binney et al. | Sep 2015 | A1 |
20150254611 | Perez | Sep 2015 | A1 |
20150259078 | Filipovic et al. | Sep 2015 | A1 |
20150286216 | Miwa | Oct 2015 | A1 |
20150314881 | Tsaliah et al. | Nov 2015 | A1 |
20150317597 | Shucker et al. | Nov 2015 | A1 |
20150332206 | Trew et al. | Nov 2015 | A1 |
20150367850 | Clarke et al. | Dec 2015 | A1 |
20150370251 | Siegel et al. | Dec 2015 | A1 |
20160001877 | Paulos | Jan 2016 | A1 |
20160009413 | Lee et al. | Jan 2016 | A1 |
20160019495 | Kolchin | Jan 2016 | A1 |
20160033966 | Farris et al. | Feb 2016 | A1 |
20160058181 | Han et al. | Mar 2016 | A1 |
20160068267 | Liu et al. | Mar 2016 | A1 |
20160085238 | Hayes | Mar 2016 | A1 |
20160104099 | Villamar | Apr 2016 | A1 |
20160104113 | Gorlin | Apr 2016 | A1 |
20160107750 | Yates | Apr 2016 | A1 |
20160114488 | Medina et al. | Apr 2016 | A1 |
20160117931 | Chan et al. | Apr 2016 | A1 |
20160125746 | Kunzi et al. | May 2016 | A1 |
20160129592 | Saboo et al. | May 2016 | A1 |
20160130015 | Caubel et al. | May 2016 | A1 |
20160132059 | Mason et al. | May 2016 | A1 |
20160144734 | Wang et al. | May 2016 | A1 |
20160144982 | Sugumaran | May 2016 | A1 |
20160180618 | Ho et al. | Jun 2016 | A1 |
20160196755 | Navot et al. | Jul 2016 | A1 |
20160196756 | Prakash et al. | Jul 2016 | A1 |
20160200438 | Bokeno et al. | Jul 2016 | A1 |
20160207627 | Hoareau et al. | Jul 2016 | A1 |
20160214717 | Silva | Jul 2016 | A1 |
20160214728 | Rossi et al. | Jul 2016 | A1 |
20160235236 | Byers et al. | Aug 2016 | A1 |
20160236778 | Takayama et al. | Aug 2016 | A1 |
20160239789 | Hanks | Aug 2016 | A1 |
20160239803 | Borley et al. | Aug 2016 | A1 |
20160257401 | Buchmueller et al. | Sep 2016 | A1 |
20160258775 | Santilli et al. | Sep 2016 | A1 |
20160266578 | Douglas et al. | Sep 2016 | A1 |
20160282126 | Watts et al. | Sep 2016 | A1 |
20160299233 | Levien et al. | Oct 2016 | A1 |
20160321503 | Zhou | Nov 2016 | A1 |
20160334229 | Ross et al. | Nov 2016 | A1 |
20160340021 | Zhang et al. | Nov 2016 | A1 |
20160364660 | Brown | Dec 2016 | A1 |
20160364679 | Cao | Dec 2016 | A1 |
20160364823 | Cao | Dec 2016 | A1 |
20160364989 | Speasl et al. | Dec 2016 | A1 |
20170011333 | Greiner et al. | Jan 2017 | A1 |
20170011340 | Gabbai | Jan 2017 | A1 |
20170032315 | Gupta et al. | Feb 2017 | A1 |
20170087999 | Miller et al. | Mar 2017 | A1 |
20170096222 | Spinelli et al. | Apr 2017 | A1 |
20170098378 | Soundararajan et al. | Apr 2017 | A1 |
20170100837 | Zevenbergen et al. | Apr 2017 | A1 |
20170101017 | Streett | Apr 2017 | A1 |
20170113352 | Lutz et al. | Apr 2017 | A1 |
20170147975 | Natarajan et al. | May 2017 | A1 |
20170154347 | Bateman | Jun 2017 | A1 |
20170164319 | Skaaksrud et al. | Jun 2017 | A1 |
20170167881 | Rander et al. | Jun 2017 | A1 |
20170193442 | Ekkel et al. | Jul 2017 | A1 |
20170199522 | Li et al. | Jul 2017 | A1 |
20170255896 | Dyke | Sep 2017 | A1 |
20170286905 | Richardson et al. | Oct 2017 | A1 |
20170300855 | Lund et al. | Oct 2017 | A1 |
20170305526 | Thomassey | Oct 2017 | A1 |
20170308098 | Yu et al. | Oct 2017 | A1 |
20170316379 | Lepek et al. | Nov 2017 | A1 |
20170330145 | Studnicka et al. | Nov 2017 | A1 |
20170345245 | Torresani et al. | Nov 2017 | A1 |
20170372256 | Kantor et al. | Dec 2017 | A1 |
20180024554 | Brady et al. | Jan 2018 | A1 |
20180088586 | Hance et al. | Mar 2018 | A1 |
20180127211 | Jarvis et al. | May 2018 | A1 |
20180137454 | Kulkarni et al. | May 2018 | A1 |
20180203464 | Yu et al. | Jul 2018 | A1 |
20190135403 | Perry et al. | May 2019 | A1 |
Number | Date | Country |
---|---|---|
2019101130 | Jan 2020 | AU |
101385059 | Mar 2009 | CN |
103224026 | Jul 2013 | CN |
203039518 | Jul 2013 | CN |
104760704 | Jul 2015 | CN |
102011086497 | May 2013 | DE |
3415436 | Dec 2018 | EP |
2692064 | Dec 1993 | FR |
2455374 | Jun 2009 | GB |
S48088255 | Oct 1973 | JP |
S56048952 | Nov 1981 | JP |
H0712088 | Jan 1995 | JP |
2004126800 | Apr 2004 | JP |
2009297449 | Dec 2009 | JP |
2010095246 | Apr 2010 | JP |
2011211025 | Oct 2011 | JP |
2012037204 | Feb 2012 | JP |
2007052246 | May 2007 | WO |
2008147484 | Dec 2008 | WO |
2008147484 | Feb 2009 | WO |
2009153588 | Dec 2009 | WO |
2010070717 | Jun 2010 | WO |
2012012752 | Jan 2012 | WO |
2012012752 | Jun 2012 | WO |
2013140085 | Sep 2013 | WO |
2013148123 | Oct 2013 | WO |
2014064431 | May 2014 | WO |
2014068982 | May 2014 | WO |
2014080409 | May 2014 | WO |
2014064431 | Jul 2014 | WO |
2017064202 | Apr 2017 | WO |
2018156991 | Aug 2018 | WO |
Entry |
---|
Andrew Amato, “Parrot Announces New Bobop Drone”, Dronelife.com, May 12, 2014; http://dronelife.com/2014/05/12/parrot- announces-new-bebop-drone/, accessed Dec. 21, 2016. |
Bullock et al., “Analysis of the Use of Digital Road Maps in Vehicle Navigation,” 1994, IEEE, p. 494-501 (Year: 1994). |
DHL Trend Research, “Self-Driving Vehicles in Logistics,” Dec. 2014, Markus Kückelhaus et al. (downloaded from http://www.dhl.com/content/dam/downloads/g0/about_us/logistics_insights/dhl_self_driving_vehicles.pdf with an archived Web version available on https://web.archive.org/web/20151018154844/http://www.dhl.com/content/dam/downloads/g0/about_us/logistics_insights/dhl_self_driving_vehicles.pdf), 39 pages. |
DHL Trend Research, “Unmanned Aerial Vehicles in Logistics: A DHL perspective on implications and use cases for the logistics industry,” 2014, Markus Kückelhaus et al., URL: http://www.dhl.com/content/dam/downloads/g0/about_us/logistics_insights/dhl_trend_report_uav.pdf with a Web Archive version available at: https://web.archive.org/web/20150923080141/http://www.dhl.com/en/about_us/logistics_insights/dhl_trend_research/uav.html, 24 pages. |
Hawas et al., “Infrastructureless Inter-Vehicular Real-Time Route Guidance,” 2008, IEEE, p. 1213-1219 (Year: 2008). |
Ivan Barajas, “Parrot Unveils New Drone and Controller,” May 14, 2014; https://blog.newegg.com/parrot-unveils-drone-controller-oculus-compatible/, accessed Dec. 21, 2016. |
Kais, Mikael et al., “An Intelligent architecture for automated transportation in our cities”, 2001 European Control Conference (ECC), Porto, Portugal, Sep. 4-7, 2001, pp. 277-282 (Year: 2001). |
Kladis et al., “Event-Based Energy Optimum Route Planning in the Context of Unmanned Aerial Vehicles for Multi-Objective Exploration Missions,” 2009, IEEE, p. 1281-1286 (Year: 2009). |
Marcus Wohlsen, “The Next Big Thing You Missed: Amazon's Delivery Drones Could Work—They Just Need Trucks,” Wired: Business, Jun. 10, 2014, URL: https://www.wired.com/2014/06/the-next-big-thing-you-missed-delivery-drones-launched-from-trucks-are-the-future-of-shipping/, 4 pages. |
Microchip Technology Inc. AT42QT1010 Data Sheet. Copyright © 2017 Microchip Technology Inc. 30 pages. |
MicroPilot home page, https://www.micropilot.com/; archived on Jan. 3, 2013 by Internet Archive, http://web.archive.org/web/2013103180733/http://www.micropilot.com/?, accessed May 2, 2016. |
Mike Murphy, “Google wants to deliver packages from self-driving trucks,” published Feb. 9, 2016, URL: https://qz.com/613277/google-wants-to-deliver-packages-from-self-driving-trucks/, 4 pages. |
Nanohawk Aeraccess: “Unmanned Aerial Vehicle—Aeraccess—Nanohawk—Webarchive 20190616”, Jun. 16, 2019 (Jun. 16, 2019), pp. 1-3, XP055759013,URL: https://web.archive.org/web/20190616235755/http://www.aeraccess-group.com/product/indoor/nanohawk [Retrieved from the Internet: Dec. 10, 2020]; the whole document. |
Navaravong et al., “Formation Reconfiguration for Mobile Robots with Network Connectivity Constraints,” 2012, IEEE, p. 18-24 (Year: 2012). |
Norman Chan, “Hands-On with Parrot's Bebop Drone Quadcopter”, tested.com, May 12, 2014; http://www.tested.com/tech/robots/461148-parrots-bebop-drone-what-you-should-know/, accessed Dec. 22, 2016. |
Parent, Michel et al., “Intelligent Transportation in Cities with CTS”, The IEEE 5th International Conference on Intelligent Transportation Systems, Sep. 3-6, 2002, Singapore, pp. 826-830 (Year: 2002). |
rchelicopterfun.com, “Parrot AF Drone Quadcopter Review”, http://www.rchelicopterfun.com/parrot-ar-drone.html; archived on Mar. 27, 2012 by Internet Archive, http://web.archive.org/web/20120327140509/http://www.rchelicopterfun.com/parrot-ar-drone.html, accessed Apr. 27, 2016. |
Sandoval, “Google patents secure rolling box to receive packages from drones,” Geekwire.com, Jan. 27, 2016, URL: http://www.geekwire.com/2016/google-pondering-drone-delivery-even-about-boxes-it-flies-to-front-doors/, 11 pages. |
Smith, Randall C., and Peter Cheeseman, “On the Representation and Estimation of Spatial Uncertainty,” The Int'l Journal of Robotics Research, vol. 5, No. 4 (Winter 1986), Copyright 1986 Massachusetts Institute of Technology, 14 pages. |
URL: https://web.archive.org/web/20160804001046/https://www.starship.xyz/, download date: Aug. 4, 2016, 21 pages. |
Wang, David. Application Report: FDC1004: Basics of Capacitive Sensing and Applications. SNOA927A—Dec. 2014—Revised Jun. 2021. Copyright © 2021 Texas Instruments Incorporated. 12 pages. |
Wikipedia, “Apollo Abort Modes”, https://en.wikipedia.org/wiki/Apollo_abort_modes; archived by Internet Archive on Aug. 16, 2012, https://web.archive.org/web/20120816115115/https://en .wikipedia.org/wiki/Apollo_abort_modes; accessed Jul. 9, 2019 (Year: 2012). |
Wikipedia, “Space Shuttle Abort Modes,” https://en.wikipedia.org/wiki/Space_Shuttle_abort_modes; archived by Internet Archive on Aug. 15, 2012, https://web.archive.org/web/20120815031221/https://en.wikipedia.org/wiki/Space_Shuttle_abort_modes; accessed Jul. 9, 2019 (Year: 2012). |
Wikipedia, “Ramer-Douglas-Peucker Algorithm,” www.wikipedia.com, Wikipedia: The Free Encyclopedia, 5 pages, Aug. 18, 2022, URL: https://en.wikipedia.org/wiki/Ramer%E2%80%93Douglas%E2%80%93Peucker_algorithm, obtained via web.archive.org on Jan. 6, 2023. |