Field of Art
The disclosure generally relates to the field of remote controllers for remote controlled vehicle, e.g., unmanned aerial vehicles, and in particular to a directional antenna in a remote controller.
Description of Art
Remote controlled or unmanned aerial vehicles, such as quadcopters, are known. Aerial vehicles continue to grow in popularity for both their commercial applications as well as recreational uses by hobbyists.
The ability of remote controlled aerial vehicles to quickly traverse space and to access places which a user cannot provides for many useful applications. However, a remote controlled aerial vehicle must, in general, maintain communicative contact with a remote controller, held by the user. Loss of connection between a remote controlled aerial vehicle and its remote controller can be potentially catastrophic. Without user control, a remote controlled aerial vehicle may crash or may otherwise be lost. Thus, the utility of the aerial vehicle is restricted by the effective communication range of the receivers and transmitters in the remote controller and aerial vehicle.
The effective communication range of the aerial vehicle may be extended by increasing the transmit power of the antennas used for communication. However, a communication system with high transmit power may require more expensive communication electronics and cause significant battery drain. Furthermore, maximum radiated power is often limited by government regulations.
Hence, there is a need to resolve these issues by extending the communication range of the aerial vehicle and remote controller.
The disclosed embodiments have advantages and features which will be more readily apparent from the detailed description, the appended claims, and the accompanying figures (or drawings). A brief introduction of the figures is below.
The Figures (FIGS.) and the following description relate to preferred embodiments by way of illustration only. It should be noted that from the following discussion, alternative embodiments of the structures and methods disclosed herein will be readily recognized as viable alternatives that may be employed without departing from the principles of what is claimed.
Reference will now be made in detail to several embodiments, examples of which are illustrated in the accompanying figures. It is noted that wherever practicable similar or like reference numbers may be used in the figures and may indicate similar or like functionality. The figures depict embodiments of the disclosed system (or method) for purposes of illustration only. One skilled in the art will readily recognize from the following description that alternative embodiments of the structures and methods illustrated herein may be employed without departing from the principles described herein.
Disclosed, by way of example embodiments, is a device communicatively coupled to an aerial vehicle. The device may be, for example, a remote controller that allows a user to control the aerial vehicle. The device includes one or more directional antennas which provides for antenna gain. Compared to the theoretical isotropic antenna, the power transmitted or received by directional antennas will be greater in a certain region of space, but weaker in another region. A directional antenna may be configured to have a main beam in which the signal power is the highest.
The aerial vehicle may detect its global position, heading, and altitude using some combination of a global positioning satellite (GPS) receiver, gyroscopes, accelerometers, a barometer, and a magnetometer. The aerial vehicle transmits this position and heading information to the communicatively coupled device. The device also may detect its own global position and heading, and estimates the position and heading of the aerial vehicle relative to itself. The device may then determines whether its antenna(s) is/are oriented so as to provide maximum (or sufficient) signal power in the direction of the aerial vehicle based on the estimated distance vector between the aerial vehicle and the device communicatively coupled with the aerial vehicle, the detected orientation of the device, and a priori information stored on the device about the radiation pattern (e.g., the far field pattern) of one or more antennas of the device.
If the current orientation of the device is not optimal, the device may display on a screen an indication that the device should be reoriented relative to the location of the aerial vehicle. This indication may include, for example, a visual indicator on a display of the aerial vehicle or an audio queue output by an electroacoustic transducer of the device. In this manner, the device coupled to the aerial vehicle may provide the user with feedback which allows the user to orient the device optimally for communication with the aerial vehicle.
If the device uses the same antenna for transmission and reception, due to the reciprocity theorem of electromagnetics, the direction in which the sensitivity of the antenna of the device to received signals is maximized is the same direction in which the transmission power is maximized. Thus, adjust the position of the antenna of the device to improve signal transmission will also improve signal reception.
A gimbal 175 may be coupled to the housing 130 of the aerial vehicle 110 through a removable coupling mechanism that mates with a reciprocal mechanism on the aerial vehicle 110. The coupling between the gimbal 175 and the aerial vehicle 110 may have mechanical and communicative capabilities. In some embodiments, the gimbal 175 may be attached or removed from the aerial vehicle 110 without the use of tools. A camera 115 may be mechanically coupled to the gimbal 175, so that the gimbal 175 steadies and controls the orientation of the camera 115. It is noted that in alternate embodiments, the camera 115 and the gimbal 175 may be an integrated configuration.
The aerial vehicle 110 may communicate with a device via a wireless network 125. The device that communicates with the aerial vehicle 110 is described herein as being a remote controller 120. However, in alternate embodiments, the device may be any other computing device capable of wireless communicating (e.g., transmitting, receiving, or both) with the aerial vehicle 110. Some or all of the description attributed herein to the remote controller 120 may also be applied to other computing devices capable of wireless communication. Other computing devices may include a device with a screen that is used to display images or video captured by the aerial vehicle but not to control the aerial vehicle, such as, a laptop, smartphone, tablet, or head-mounted display.
In one embodiment, the wireless network 125 may be a long range Wi-Fi system. It also may include or be another wireless communication system, for example, one based on long term evolution (LTE), 3G, 4G, or 5G mobile communication standards. In some embodiments, the wireless network 125 consists of a single channel and the aerial vehicle 110 and the remote controller 120 implement a half-duplex system. In an alternate embodiment, the wireless network 125 includes two channels: a unidirectional channel used for communication of control information from the remote controller 120 to the aerial vehicle 110 and a separate unidirectional channel used for video downlink from the aerial vehicle 110 to the remote controller 120 (or to another device, such as a video receiver where direct video connection may be desired). Alternate wireless network configurations may also be used.
The remote controller 120 in this example includes a first control panel 150, a second control panel 155, an ignition button 160, a return button 165, and a screen (or display) 170. The first control panel 150 may be used to control “up-down” direction (e.g. lift and landing) of the aerial vehicle 110. The second control panel 155 may be used to control “forward-reverse” or may control the direction of the aerial vehicle 110. In alternate embodiments, the control panels 150, 155 are mapped to different directions for the aerial vehicle 110. Each control panel 150, 155 may be structurally configured as a joystick controller and/or touch pad controller. The ignition button 160 may be used to start the rotary assembly (e.g., start the propellers 145). The return button 165 may be used to override the controls of the remote controller 120 and transmit instructions to the aerial vehicle 110 to autonomously return to a predefined location. The ignition button 260 and the return button 265 may be mechanical and/or solid state press sensitive buttons.
In addition, each button may be illuminated with one or more light emitting diodes (LEDs) to provide additional details. For example a LED may switch from one visual state to another to indicate with respect to the ignition button 160 whether the aerial vehicle 110 is ready to fly (e.g., lit green) or not (e.g., lit red) or whether the aerial vehicle 110 is now in an override mode on return path (e.g., lit yellow) or not (e.g., lit red). It also is noted that the remote controller 120 may include other dedicated hardware buttons and switches and those buttons and switches may be solid state buttons and switches. For example, a button or switch may be configured to allow for triggering a signal to the aerial vehicle 110 to immediately execute a landing operation.
The remote controller 120 may also include hardware buttons or other controls that control the gimbal 175 or camera 115. The remote controller 120 may allow it's user to change the preferred orientation of the camera 115. In some embodiments, the preferred orientation of the camera 115 may be set relative to the angle of the aerial vehicle 110. In another embodiment, the preferred orientation of the camera 115 may be set relative to the ground. The remote controller 120 may also transmit commands to the aerial vehicle 110 which are routed to the camera 115 through the gimbal 175 to take a picture, record a video, change a picture or video setting, and the like.
The remote controller 120 also may include a screen 170 which provides for visual display. The screen 170 may be a touch sensitive screen. The screen 170 also may be, for example, a liquid crystal display (LCD), an LED display, an organic LED (OLED) display, or a plasma screen. The screen 170 may allow for display of information related to the remote controller 120, such as menus for configuring the remote controller 120 or remotely configuring the aerial vehicle 110. The screen 170 also may display images or video captured from the camera 115 coupled with the aerial vehicle 110, wherein the images and video are transmitted to the remote controller 120 via the wireless network 125. The video content displayed on the screen 170 may be a live feed of the video or a portion of the video captured by the camera 115. It is noted that the video content may be displayed on the screen 170 within a short time (e.g., within fractions of a second) of being captured by the camera 115. The delay between the video being captured by the camera 115 and being displayed on the screen 170 may be instantaneous or nearly instantaneous in terms of human perceptual quality.
The video may be overlaid and/or augmented with other data from the aerial vehicle 110 such as the telemetric data from a telemetric subsystem of the aerial vehicle 110. The telemetric subsystem may include navigational components, such as a gyroscope, an accelerometer, a compass, a global positioning system (GPS) and/or a barometric sensor. In one example embodiment, the aerial vehicle 110 may incorporate the telemetric data with video that is transmitted back to the remote controller 120 in real time. The received telemetric data may be extracted from the video data stream and incorporated into predefine templates for display with the video on the screen 170 of the remote controller 120. The telemetric data also may be transmitted separate from the video from the aerial vehicle 110 to the remote controller 120. Synchronization methods such as time and/or location information may be used to synchronize the telemetric data with the video at the remote controller 120. This example configuration may allow a user of the remote controller 120 to see where the aerial vehicle 110 is flying along with corresponding telemetric data associated with the aerial vehicle 110 at that point in the flight. Further, if the user is not interested in telemetric data being displayed real-time, the data may still be received and later applied for playback with the templates applied to the video.
The predefine templates may correspond with “gauges” that provide a visual representation of speed, altitude, and charts, e.g., as a speedometer, altitude chart, and a terrain map. The populated templates, which may appear as gauges on a screen 170 of the remote controller 120, may further be shared, e.g., via social media, and or saved for later retrieval and use. For example, a user may share a gauge with another user by selecting a gauge (or a set of gauges) for export. Export may be initiated by clicking the appropriate export button, or a drag and drop of the gauge(s). A file with a predefined extension may be created at the desired location.
The remote controller 120 shown in
The remote controller 120 may include a screen 170 and a speaker 320 (e.g., an electroacoustic transducer) for providing output to a user. The speaker 320 may output sound from a video as it is displayed on the screen 170. The video may be received from the aerial vehicle 110 via the wireless network 125. The speaker 320 may also output sounds responsive to the user pressing a button or as an alert to the user. For example, the speaker may output an alert when the battery of the aerial vehicle 110 is nearly depleted, when an error is detected on the aerial vehicle 110 (e.g., a mechanical malfunction, a software error, an electronic malfunction, or a combination thereof), when the signal strength between the aerial vehicle 110 and the remote controller 120 is weak, when the antenna of the remote controller 120 is not oriented correctly, and/or when the wireless connection with the aerial vehicle 110 is lost.
The remote controller 120 also includes user input devices. Specifically, the remote controller 120 may include a first control panel 150, a second control panel 155, an ignition button 160, a return button 165, and a power button 310. The first control panel 150 and the second control panel, 155 may be joystick controllers for controlling the velocity and orientation of the aerial vehicle 110. The power button 310 may toggle the power of the remote controller 120 or toggle the power of the aerial vehicle 110. In some embodiments, the screen 170 may be a touch screen and thus can receive user inputs as well.
The remote controller 120 may contain one or more internal directional antennas (not shown in
The radiation patterns of the one or more directional antenna of the remote controller 120 are discussed herein with respect to the standard Cartesian coordinate system in where the x-y plane is the horizontal plane and where the z-axis is in the upward vertical direction. The one or more directional antenna of the remote controller 120 may be configured such that the main beam direction (i.e., the direction of greatest radiated energy) of each antenna is horizontal when the remote controller 120 is held horizontally. In some embodiments, the main beam directions of the antennas may be directed above the horizontal plane.
Turning now to
The radiation direction vector b 410 is defined herein as a unit vector in the direction of the radiation of the one or more antennas of the remote controller 120. In general, the radiation direction vector b 410 is the direction in which the one or more antennas of the remote controller 120 best transmit signals to and receive signals from the aerial vehicle 110. If the remote controller 120 has a single antenna or if every antenna has the same main beam axis, the direction of b 410 may be the direction of the main beam axis. The main beam axis of a directional antenna is the direction of maximum gain of the direction antenna (i.e., the boresight direction of the antenna). In embodiments where the remote controller 120 includes more than one antenna, the radiation direction b 410 may be the average of the beam axis directions of each antenna. The radiation direction b 410 may also be configured to be the optimal direction for communication using the antennas.
The second vector, v 420, is the unit vector orthogonal to the radiation direction vector b 410 and parallel with the horizontal side of the screen 170. In the normal orientation of the remote controller 120 (i.e., without roll rotation), v 420 is on the x-y plane. In an embodiment where the remote controller 120 lacks a screen 170, v 420 may be synonymously defined as being on the horizontal plane (i.e., the x-y plane) when the remote controller 120 is held at an orientation without a roll rotation. Finally, the third vector, w 430, is the vector cross product of b 410 and v 420. Thus, w 430 is orthogonal to both b 410 and v 420, and when v 420 is on the x-y plane (i.e., when the remote controller 120 has no roll rotation), w 430 is on the same vertical plane as b 410. If the remote controller 120 is held without a roll rotation such that the radiation direction vector b 410 lies on the x axis, then b 410, v 420, and w 430 lie on the x, y, and z axes, respectively. These vectors {b,w,v} are defined based on the radiation pattern of one or more antennas of the remote controller 120, and, therefore, based on the orientation of the remote controller 120. Thus, any reorientation of the remote controller 120 will also change two or more of the vectors b 410, v 420, and w 430.
The processing system 510 may be configured to provide the electronic processing infrastructure to execute firmware and software comprised of instructions. The processing system 510 may include one or more hardware processors.
The navigation system 520 may include electronics, controls, and interfaces for navigation instrumentation of the remote controller 120. The navigation system 520 may include a global position system (GPS) receiver, a compass (e.g., a magnetometer), gyroscopes, accelerometers, a barometer, or some combination thereof. The GPS receiver, accelerometers, gyroscopes, and compass may be used to track the location, motion, and orientation of the remote controller 120.
The I/O system 530 may include the input and output interfaces and electronic couplings to interface with devices that allow for transfer of information into or out of the remote controller 120. For example, the I/O system 530 may include a physical interface such as a universal serial bus (USB) or a media card (e.g., secure digital (SD)) slot. The I/O system 530 also may be associated with the communication subsystem 570 to include a wireless interface such as Bluetooth™. In addition, it is noted that in one example embodiment, the aerial vehicle 110 may use long range Wi-Fi radio within the communication system 570, but may also use a second Wi-Fi radio or cellular data radio (as a part of the I/O system 530) for connection to other wireless data enabled devices, for example, smart phones, tablets, laptop or desktop computers, and wireless internet access points. Moreover, the I/O system 530 may also include other wireless interfaces, e.g., Bluetooth, for communicatively coupling devices that are similarly wirelessly enabled for short range communications.
The display system 540 may be configured to provide an interface, electronics, and display drivers for one or more display screens (e.g., the screen 170 of the remote controller 120). The audio/visual system 550 may include the interfaces, electronics, and drivers for an audio output (e.g., headphone jack or speakers) as well as visual indicators (e.g., LED lighting associated with, for example, the buttons 160, 165 of the remote controller 120). The control system 560 may include electronic and control logic and firmware for operation with the control panels 150, 155.
The communication system 570 may include electronics, firmware, and interfaces for communications. The communications system 570 may include one or more of wireless communication mechanisms such as Wi-Fi (short and long range), LTE, 3G/4G/5G, and the like. The communication system 570 also may include wired communication mechanisms such as Ethernet, USB, and high-definition multimedia interface (HDMI). The communication system 570 may include an antenna system 575 coupled to a receiver, transmitter, or transceiver. The antenna system 575 may include one or more antennas. A transceiver of the communication system 570 may transmit and receive data with the antenna system 575 to communicate over the wireless network 125. Prior to transmitting data from the remote controller 120, the transceiver may process the data by performing encryption, scrambling, forward error correction coding (FEC), lossless compression, lossy compression, or some combination thereof. Similarly, the transceiver may process data received over the wireless network 125 by performing decryption, unscrambling, error correction decoding, decompression, or some combination thereof.
The power system 580 may include electronics, firmware and interfaces for providing power to the system. The power system 580 may include direct current (DC) power sources (e.g., batteries), but also may be configured for alternating current (AC) power sources. The power system 580 also may include power management processes for extending DC power source lifespan. It is noted that in some embodiments, the power system 580 may be comprised of a power management integrated circuit and a low power microprocessor for power regulation. The microprocessor, in such embodiments, may be configured to provide very low power states to preserve battery, and have the ability to wake from low power states in response to such events as a button press or an on-board sensor (e.g., a hall sensor) trigger.
As described in greater detail below, the camera 115 may include sensors 640 to capture metadata associated with video data, such as timing data, motion data, speed data, acceleration data, altitude data, GPS data, and the like. In a particular embodiment, location and/or time centric metadata (e.g., geographic location, time, and/or velocity) may be incorporated into a media file together with the captured content in order to track the location of the camera 115 over time. This metadata may be captured by the camera 115 itself or by another device (e.g., a mobile phone or the aerial vehicle 110) proximate to the camera 115. In one embodiment, the metadata may be incorporated with the content stream by the camera 115 as image content is being captured. In another embodiment, a metadata file separate from the video file may be captured (by the same capture device or a different capture device) and the two separate files may be combined or otherwise processed together in post-processing. It is noted that these sensors may be in addition to the sensors 640 of the camera 115. In embodiments in which the camera 115 is integrated into the aerial vehicle 110, the camera 115 need not have separate individual sensors, but rather may rely upon the sensors integrated with the aerial vehicle 110.
Referring now to the details of
The lens 612 may be, for example, a normal lens, a wide angle lens, a hemispherical lens, or a hyper-hemispherical lens that focuses light entering the lens to the image sensor 614 which captures images and/or video frames. The image sensor 614 may capture high-definition images having a resolution of, for example, 720p, 1080p, 4 k, or higher. For video, the image sensor 614 may capture video at frame rates of, for example, 30 frames per second, 60 frames per second, or higher. The image processor 616 may perform one or more image processing functions on the captured images or video. For example, the image processor 616 may perform a Bayer transformation, demosaicing, noise reduction, image sharpening, image stabilization, rolling shutter artifact reduction, color space conversion, compression, or other in-camera processing functions. Processed images and video may be temporarily or persistently stored to system memory 630 and/or to a non-volatile storage, which may be in the form of internal storage or an external memory card.
An input/output (I/O) interface 660 may transmit and receive data from various external devices. For example, the I/O interface 660 may facilitate the receiving or transmitting of video or audio information through an I/O port. Examples of I/O ports or interfaces include USB ports, HDMI ports, Ethernet ports, audio ports, and the like. Furthermore, embodiments of the I/O interface 660 may include wireless ports that can accommodate wireless connections. Examples of wireless ports include Bluetooth, Wireless USB, Near Field Communication (NFC), and the like. The I/O interface 660 may also include an interface to synchronize the camera 115 with other cameras or with other external devices, such as a remote control, a second camera, a smartphone, a client device, or a video server.
A control/display subsystem 670 may include various control and display components associated with operation of the camera 115 including, for example, LED lights, a display, buttons, microphones, speakers, and the like. The audio subsystem 650 may include, for example, one or more microphones and one or more audio processors to capture and process audio data correlated with video capture. In one embodiment, the audio subsystem 650 may include a microphone array having two or more microphones arranged to obtain directional audio signals.
The sensors 640 capture various metadata concurrently with, or separately from, video capture. For example, the sensors 640 may capture time-stamped location information based on a global positioning system (GPS) sensor, and/or an altimeter. Other sensors 640 may be used to detect and capture the orientation of the camera 115 including, for example, an orientation sensor, an accelerometer, a gyroscope, or a magnetometer. Sensor data captured from various sensors on the aerial vehicle 110 may be processed to generate other types of metadata. For example, sensor data from the accelerometer may be used to generate motion metadata which may include velocity and/or acceleration vectors representative of motion of the camera 115. Furthermore, sensor data from the aerial vehicle 110 and/or the gimbal 175 may be used to generate orientation metadata describing the orientation of the camera 115. Sensor data from a GPS receiver may provide GPS coordinates identifying the location of the camera 115, and an altimeter may measure the altitude of the camera 115. In one embodiment, the sensors 640 may be rigidly coupled to the camera 115 such that any motion, orientation, or change in location affecting the camera 115 also affects the sensors 640. The sensors 640 furthermore may associates a time stamp representing when the data was captured by each sensor. In one embodiment, the sensors 640 may automatically begin collecting sensor metadata when the camera 115 begins recording a video.
The map 710 may include a user indicator 711 and a vehicle indicator 712, which indicate the geographic position of the remote controller 120 and the aerial vehicle 110, respectively. The map 710 may also display a vector map which displays geographic features such as bodies of water, rivers, cliffs, inclines, mountains, or other geographical features. The vector map may be a two-dimensional (2D) map or a three-dimensional (3D) map. In some embodiments, the 3D vector map is displayed as a topographical map. The vector map may also display manmade features such as roads, hiking paths, the footprint of buildings, 3D models of buildings, and notable landmarks. In some embodiments, the map 710 also displays overhead imagery, such as satellite imagery.
The relative position of the elements displayed on the map 710 may be indicative of the location of corresponding geographical and manmade features. The map 710 may display a subsection of a vector map or image stored on the remote controller 120. In some embodiments, the remote controller 120 may receive maps based on its detected GPS location from a remote server through a wireless network.
The information panel 720 displays information relating to the aerial vehicle 110. The information panel 720 may include a remaining flight time indicator 721, an altitude indicator 722, a horizontal distance indicator 723, and a speedometer 724. The information displayed in the information panel 720 may be received from the aerial vehicle 110 via the wireless network 125. The remaining flight time indicator 721 may display an estimate of the time until the battery of the aerial vehicle 110 is depleted based on the current charge of the battery of the aerial vehicle 110 and/or the logged flight path of the aerial vehicle 110. The altitude indicator 722 may display the altitude of the aerial vehicle 110, where altitude is measured in units of distance (e.g., feet, yards, meters, etc.). The displayed altitude may be relative to the remote controller 120, the lift-off location, the ground directly underneath the aerial vehicle 110, a user defined altitude, or an absolute altitude (e.g., mean sea level).
The distance indicator 723 may display either the displacement of the aerial vehicle 110 from some reference position or a total path length. In some embodiments, the distance indicator 723 may display a displacement of the aerial vehicle 110 from the remote controller 120, the lift-off location, or a user defined position. In alternate embodiments, the distance indicator 723 may display a total path length. The path for which the path length is defined may be the flight path that the aerial vehicle 110 has flown, a programmed flight path which the aerial vehicle 110 is configured to automatically follow, or an automatic return to home path. The displacement or the path length displayed by the distance indicator 723 may be the total displacement or path length or may be the horizontal component of the displacement or path length. The speedometer 724 may display the speed of the aerial vehicle 110. The measurement of the speed may be based on the horizontal component of the velocity of the aerial vehicle 110 or may be based on the aerial vehicle's total velocity.
The antenna direction indicator 730 may be a visual indicator that displays the relative position of the aerial vehicle 110. The antenna direction indicator 730 may include a vehicle icon 731, a radiation direction indicator 732, and an orientation adjustment arrow 733. The antenna direction indicator 730 may display an indication of the direction of the displacement vector between the remote controller 120 and the aerial vehicle 110. Herein, this displacement vector is denoted as d. The vertical component of the direction of d (i.e., the angle that the displacement vector d makes with the horizontal plane) is denoted as θβ and the horizontal component of the direction of d relative to some reference direction is denoted as θα. The antenna direction indicator 730 also displays the direction of the radiation pattern of the antenna system 575 of the remote controller 120. As described in conjunction with
The position of the vehicle icon 731 on the circle of the antenna direction indicator 730 may be based on θα, the horizontal component of the displacement vector d between the remote controller 120 and the aerial vehicle 110. This angle θα may be relative to a fixed orientation, such as polar north or magnetic north. For example, the vehicle icon 731 is displayed at a 135° position on the circle in
If the remote controller 120 completely loses its connection with the aerial vehicle 110, the vehicle icon 731 may remain at the same position. In embodiments in which the aerial vehicle 110 is configured with automatic return to home behavior, the remote controller 120 may predict the position of the aerial vehicle 110 based on the return to home path. The vehicle icon 731 may be displayed on the circle of the antenna direction indicator 730 based on this predicted position to assist the user in restoring the connection with the aerial vehicle 110.
The radiation direction indicator 732 may be a wedge which indicates the horizontal direction φα of the radiation pattern of the antenna system 575 of the remote controller 120. The orientation of the center of the wedge may be based on the horizontal direction φα of the radiation direction vector b 410, e.g., as shown in
In some embodiments, the arc length of the radiation direction indicator 732 may be based on an arc threshold Tα. Tα is a threshold for the difference between the horizontal direction φα of the radiation direction vector b 410 and the horizontal direction θα of the displacement vector d. The arc that the radiation direction indicator 732 composes may be the range of angles between φα−Tα and φα+Tα. Tα may be based on the beamwidth of the main lobes of the antennas of the antenna system 575. The beamwidth may be the angle between the points on the main lobe having half-power (−3 dB) relative to the maximum gain of the antennas of the antenna system 575. In other embodiments, the arc threshold Tα may be based on the range of yaw angles for which the gain of the remote controller 120 will be sufficient for communication between the remote controller 120 and the aerial vehicle 110. The gain required for communication, and thus the arc threshold Tα, may be based on the distance between the aerial vehicle 110 and the remote controller 120 and may also be based on the signal attenuation as a function of distance. Accordingly, in such an embodiment, the arc threshold Tα may decrease as the displacement between the aerial vehicle 110 and the remote controller 120 increases.
In some embodiments, the arc threshold Tα may be based on the vertical component of the beam direction φβ relative to vertical component of the displacement direction θβ. That is, the arc length of the radiation direction indicator 732 may be based on the angle between the radiation direction vector b 410 and the horizontal plane relative to the vertical displacement between the remote controller 120 and the aerial vehicle 110. In general, the arc threshold Tα is largest when θβ=φβ.
In some embodiments, the arc threshold Tα may also be based on the noise detected on the wireless channel used for communication, where higher noise corresponds to a smaller value of the arc threshold Tα. The noise may be indirectly detected based on received control characters from the aerial vehicle 110. For example, the arc threshold Tα may be decreased when the remote controller 120 receives negative-acknowledge characters (NACKs) or fails to receive acknowledge characters (ACKs) from the aerial vehicle 110. The arc threshold Tα may also be decreased based on a high bit error rate (BER) detected by an error-correcting decoder of the communication system 570 of the remote controller 120 or similar channel quality indicators. In alternate embodiments, the arc length of the radiation direction indicator 732 may be invariant.
The antenna direction indicator 730 also may include an orientation adjustment arrow 733. The orientation adjustment arrow 733 may indicate the direction of rotation (e.g., clockwise or counterclockwise), in the horizontal plane, for the user to rotate the remote controller 120 in order to increase the signal power received at the aerial vehicle 110. The antenna direction indicator 730 may indicate the direction of rotation of the remote controller 120 that requires the smallest angular adjustment to align the horizontal component of the radiation direction vector b 410 of the remote controller 120 with that of the displacement vector d between the remote controller 120 and the aerial vehicle 110. For example, in
The video panel 740 may display video received from the aerial vehicle 110. A camera 115 coupled to the aerial vehicle 110 may capture video and transmit it (e.g., via the aerial vehicle 110) through the wireless network 125. The antenna direction indicator 730, the map 710, or elements thereof may be partially transparent to minimize the obstruction of the video. The video panel 740 may include a recording indicator 741 which indicates that video is being recorded by the camera 115 and displays the length (e.g., in seconds) of the currently recording video.
The user interface 700 illustrated in
Depending on the radiation pattern of the antenna system 575, the low pitch of the remote controller 120 may negatively impact the power of signals transmitted between the remote controller 120 and the aerial vehicle 110. In some embodiments, such as when the radiation pattern of the antenna system 575 of the remote controller 120 has a thin main lobe 450 as in
The communication system 570 of the remote controller 120 receives 810 position information for the aerial vehicle 110. The aerial vehicle 110 may detect the position information based on sensor data collected by some combination of a GPS receiver, an electronic compass (e.g., a magnetometer), a barometer, and an inertial measurement unit (IMU) on the aerial vehicle 110. The aerial vehicle 110 may transmit the position information to the remote controller 120 through the wireless network 125. The received position information may include absolute position information. For example, the position information may specify the altitude, latitude, and longitude of the aerial vehicle 110. The remote controller 120 may also receive an indication of the some combination of the aerial vehicle's orientation (e.g., a pitch, roll, and yaw), velocity, acceleration, angular velocity, and angular acceleration. In some embodiments, receiving 810 position information for the aerial vehicle 110 includes extrapolating a future position for the aerial vehicle 110 based on previously received position, speed, and acceleration data.
The navigation system 520 of the remote controller 120 detects 820 the position of the remote controller 120. The remote controller 120 then estimates 840 the direction of displacement between the remote controller 120 and the aerial vehicle 110. In some embodiments, the remote controller 120 may estimates 840 the direction of displacement by estimating the displacement vector d by subtracting the position of the remote controller 120 from the position of the aerial vehicle 110. The direction of displacement may be the direction of horizontal displacement θα. The direction of displacement may also include the direction of vertical displacement θβ.
The navigation system 520 of the remote controller 120 also detects 830 the orientation of the remote controller 120. Detecting the orientation of the remote controller 120 may include estimating the yaw, pitch, and/or roll of the remote controller 120 based on sensor data from a GPS receiver, gyroscopes, accelerometers, and/or a magnetometer of the navigation system 520.
The radiation direction vector b 410 of the antenna system 575 of the remote controller 120 is estimated 850 based on the detected 830 orientation of the remote controller 120. In some embodiments, the direction of the radiation direction vector b 410 is estimated 850 by applying a rotational transform to a vector representing a reference radiation direction. For example, b0 may be a vector representing the radiation direction when the remote controller 120 has no yaw, pitch, or roll rotation relative to some reference rotation. The radiation direction vector b 410 may be given by b=R(α,β,γ)×b0 where α, β, and γ are the yaw, pitch, and roll angles estimated for the remote controller 120 and R(α,β,γ) is a rotation matrix, as conventionally defined. The reference radiation direction vector b0 may be a constant stored on the remote controller 120 and may be configured based on a mathematical model of the radiation pattern of the antenna system 575 or empirical methods. Alternately, the reference radiation direction vector b0 may be variable and may be determined based on a stored model of the radiation pattern of the antenna system 575. In embodiments in which the antenna system 575 may transmit signals at multiple wavelengths, the reference radiation direction vector b0 may be based on the wavelength of the signal being transmitted or received. Additionally, as discussed further below, the reference beam axis vector b0 may be based on the rotation of the hinge 350 connecting the first section 330 and the second section 340 of the remote controller 120.
The direction of displacement d between the remote controller 120 and the aerial vehicle 110 and the radiation direction vector b 410 may be used to determine 860 whether the remote controller 120 is correctly oriented. This determination may be made by comparing the angle between d and b 410 to a threshold value. For example, determining 860 whether the remote controller 120 is correctly oriented may comprise determining whether the absolute angular difference between θα and φα is less than Tα where θα is the direction of the horizontal component of the displacement vector d, φα is the direction of the horizontal component of the radiation direction vector b 410, and Tα is a threshold arc length.
If the orientation of the radiation direction vector b 410 is within an acceptable range, the remote controller 120 may display 870 a user interface (e.g., user interface 700) with an indication that the orientation of the remote controller 120 is correct. In the case of the user interface 700, this is indicated by displaying the vehicle icon 731 inside of the radiation direction indicator 732. In some embodiments, the radiation direction indicator may change visual characteristics, e.g., color and/or visual patterns, based on the determination of whether the remote controller 120 is correctly oriented. For example, the radiation direction indicator 732 may be a first color (e.g., green) when the remote controller 120 is correctly oriented and a second color (e.g., red) when it is not. The colors may also have varying gradients based on how close the orientation of the remote controller 120 is to ideal. For example, the radiation direction indicator 732 may be displayed as red when not oriented correctly (e.g., a large orientation difference), pink when closer to being oriented correctly, light green when nearly oriented correctly, to dark green when correctly oriented.
Further, it is noted that other indicators may be used, e.g., audio or tactile feedback. For example, in the case of audio, the remote controller 120 may output no audio when the remote controller 120 is correctly oriented and may outputs audio beeps of different duration and/or frequency played when it is not. The frequency, duration, or loudness of the beeps may correspond to the magnitude of the orientation error of the remote controller 120. By way of example, in the case of tactile feedback, a touch sensitive surface may produce no vibration when the remote controller 120 is correctly oriented, but may produce vibrations of different duration and/or frequency (e.g., with one or more haptic actuators) when it is not. The frequency or duration of the vibrations may correspond to the magnitude of the orientation error of the remote controller 120.
If, on the other hand, the determination is made that the remote controller 120 is not correctly oriented, an indication that the user should alter the orientation of the remote controller 120 may be displayed 880. In the case of user interface 700, this may be indicated by displaying the vehicle icon 731 outside of the radiation direction indicator 732 and/or by displaying a tilt up indicator 750 or a tilt down indicator.
The feedline 960 may carry a signal from the transceiver 950 to the first antenna 910 when the antenna system 575 is transmitting a signal, and may carry a signal from the first antenna 910 to the transceiver 950 when the antenna system 575 is receiving a signal. The feedline 960 may be a transmission line, such as a coaxial cable. The tubular conducting shield of the coaxial cable may couple to the ground plane 914. The inner conductor of the coaxial cable may couple to the probe feed 913. In some embodiments, the inner conductor of the coaxial cable is the probe feed 913. In some embodiments, a matching circuit is coupled between the feedline 960 and the first antenna 910 for impedance matching.
The probe feed 913 may be a wire perpendicular to the dielectric layer 912 which couples electromagnetic energy in or out of the patch 911. The patch 911 may be a thin, rectangular conductor (e.g., metal). The probe feed 913 may couple to the patch 911 at a point that is centered along the short axis of the patch 911 and off-center along the long axis of the patch 911.
The dielectric layer 912 may separate the patch 911 from the ground plane 914. The dielectric layer 912 may be a ceramic substrate. The ground plane 914 may be a thin sheet of conductive material (e.g., metal).
In some embodiments, the first antenna 910 and the second antenna 920 are tilted away from the x-axis (along the x-y plane).
Changing the relative orientations of the first section 330 and the second section 340 of the remote controller 120 also may affect the axes of maximum gain for the antennas 910, 920. Compared to horizontal position 1000, the axes of maximum gain for the vertical position 1100 may be higher (i.e., have a larger z-component) for both the first antenna 910 and second antenna 920. Accordingly, the radiation direction vector b 1310 for the vertical position 1100 may be directed higher as well. Thus, the angle φα 1320 that the radiation direction vector b 1310 makes with the horizontal (x-y) plane when in the remote controller 120 vertical position 1100 may be higher than the angle φα 1220 when in the remote controller 120 is in the horizontal position 1000.
In some embodiments, the remote controller 120 includes sensors which may detect the rotation of the hinge 350. The remote controller 120 may store a model of the radiation patterns for the antenna system 575 mapped to the orientation of the hinge 350. Based on the detected rotation of the hinge 350, the remote controller 120 may determine the radiation direction vector b 410, the arc threshold Tα, and/or the tilt up or tilt down thresholds Tβ1 and Tβ2. Thus, the display of antenna direction indicator 730 of the user interface 700 may be based, in part, on the detected orientation of the hinge 350.
The disclosed configuration describes a system and method for displaying a user interface 700 on a remote controller 120 to assist a user in correctly orienting the remote controller 120 for optimal communication with an aerial vehicle 110. Position information may be received by the remote controller 120 from the aerial vehicle 110. The remote controller 120 may detect its own position and orientation. Based on the orientation of the remote controller 120 and the relative position of the remote controller 120 and the aerial vehicle 110, the remote controller 120 displays an indication to the user to assist the user in orienting the remote controller 120 so that one or more directional antennas of the remote controller 120 are oriented for a high transmit power.
By way of example, the remote controller 120 may include two antennas 910, 920. These antennas 910, 920 may be ceramic patch antennas. The antennas 910, 920 may be oriented 30° apart from each other for transmit diversity. The remote controller 120 may also include a screen 170 which is attached to the rest of the remote controller 120 with a hinge 350. The orientation of the screen 170 may affect the radiation pattern of the remote controller 120. The indication in the user interface 700 that assists the user in orienting the remote controller 120 may be based, in part, on the detected rotation of the hinge 350 to account for this effect on the radiation pattern.
Throughout this specification, plural instances may implement components, operations, or structures described as a single instance. Although individual operations of one or more methods are illustrated and described as separate operations, one or more of the individual operations may be performed concurrently, and nothing requires that the operations be performed in the order illustrated. Structures and functionality presented as separate components in example configurations may be implemented as a combined structure or component. Similarly, structures and functionality presented as a single component may be implemented as separate components. These and other variations, modifications, additions, and improvements fall within the scope of the subject matter herein.
Unless specifically stated otherwise, discussions herein using words such as “processing,” “computing,” “calculating,” “determining,” “presenting,” “displaying,” or the like may refer to actions or processes of a machine (e.g., a computer) that manipulates or transforms data represented as physical (e.g., electronic, magnetic, or optical) quantities within one or more memories (e.g., volatile memory, non-volatile memory, or a combination thereof), registers, or other machine components that receive, store, transmit, or display information.
As used herein any reference to “one embodiment” or “an embodiment” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
Some embodiments may be described using the expression “coupled” and “connected” along with their derivatives. For example, some embodiments may be described using the term “coupled” to indicate that two or more elements are in direct physical or electrical contact. The term “coupled,” however, may also mean that two or more elements are not in direct contact with each other, but yet still co-operate or interact with each other. The embodiments are not limited in this context.
As used herein, the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having” or any other variation thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Further, unless expressly stated to the contrary, “or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present).
In addition, use of the “a” or “an” are employed to describe elements and components of the embodiments herein. This is done merely for convenience and to give a general sense of the invention. This description should be read to include one or at least one and the singular also includes the plural unless it is obvious that it is meant otherwise.
Upon reading this disclosure, those of skill in the art will appreciate still additional alternative structural and functional designs for the disclosed remote controller, the user interface thereof, and associated systems. Thus, while particular embodiments and applications have been illustrated and described, it is to be understood that the disclosed embodiments are not limited to the precise construction and components disclosed herein. Various modifications, changes and variations, which will be apparent to those skilled in the art, may be made in the arrangement, operation and details of the method and apparatus disclosed herein without departing from the spirit and scope defined in the appended claims.
This application claims the benefit of U.S. Provisional Patent Application No. 62/267,176 filed on Dec. 14, 2015, the content of which is incorporated by reference in its entirety herein.
Number | Date | Country | |
---|---|---|---|
62267176 | Dec 2015 | US |