The disclosure relates generally to aircraft camera systems and viewing systems to allow passengers within the aircraft to enjoy views of the ground below and views of the aircraft in its environment.
Aircraft passengers enjoy looking out the cabin windows and also enjoy viewing the aircraft's en route flight progress on display monitors. These are standard aircraft features. Some aircraft are additionally outfitted with external fore and aft video cameras to provide passengers with views of the runway during takeoff and landing and views looking downward while the aircraft is in flight. Typically these video cameras capture images that are displayed on display monitors within the cabin, often the same monitors used to display en route flight progress.
Unfortunately, viewing perspectives of both cabin windows and external video monitors are limited; and when the aircraft is flying at night or when the aircraft is over weather, there may be little or nothing of interest to see. Moreover, when the aircraft is at cruising altitude, the ground is so far away that very little detail can be seen.
Increasing resolution of the in-cabin display monitors is no practical solution, because camera resolution is the limiting factor. It is not commercially practical to constantly upgrade camera resolution to match the display resolution, because each camera upgrade will typically requires an extensive regulatory process.
Monitor resolution actually presents another problem, as well. Passengers have come to expect ever higher and higher display resolutions, to match the resolutions produced by the latest mobile phones and high definition home entertainment systems. With each iterative resolution increase in consumer electronic displays, the images from aircraft-mounted legacy video cameras just seem to get worse and worse by comparison.
Disclosed here is a new approach, which gives aircraft passengers a much higher resolution and more flexible view of both the ground below and the aircraft in its environment. Instead of using aircraft-mounted video cameras, the disclosed system uses a virtual reality image synthesis system coupled to an on-board database containing previously captured, photo-realistic images of the ground below the aircraft's current position.
The virtual reality system offers passengers the ability to view the ground and the aircraft's environment from a full range of different perspectives, including a synthesized composite view showing a synthesized rendering of the actual aircraft as seen from the outside looking in—as the aircraft would appear to another aircraft flying beside, above or below it.
One advantage of the virtual reality system stems from the fact that the field of view provided to the passenger is physically decoupled from the actual altitude at which the aircraft is flying. For example, when flying at a cruising altitude of 40,000 feet, the virtual reality system can be programmed to supply an image of the ground, as it would appear at an altitude of 5000 feet. In presenting this perspective, the virtual reality system takes the aircraft's actual position and ground speed into account, so that objects directly beneath the physical aircraft will always be presented directly beneath the aircraft as rendered in virtual reality space.
This ability to decouple cruising altitude from the rendered virtual image is something quite foreign to aircraft-mounted video camera systems, where the perspective of the video camera is always tied to the actual altitude of the plane because the camera and its lens system are physically attached to the underside of the aircraft.
In the disclosed system, if desired, the passenger can control the virtual reality system to request a view showing where the aircraft was previously located, or where the aircraft will (or could be) located in the future.
In one aspect, the disclosed virtual camera system supplies terrain views to occupants of an aircraft, using a virtual reality image generator disposed on the aircraft. The image generator generates a terrain display image corresponding to a perspective based on supplied position data. The virtual camera system further includes an aircraft position sensor disposed on the aircraft and coupled to supply position data to the virtual reality image generator. A display device disposed on the aircraft is coupled to the virtual reality image generator to reproduce the terrain display image for viewing by passengers within the aircraft.
The virtual reality image generator includes a processor that may be programmed to generate the terrain display image using actual aircraft latitude and longitude relative to the Earth and further using a synthetic altitude that differs from the actual aircraft altitude. In some embodiments, the synthetic altitude is used above actual aircraft altitudes of a predetermined threshold. When below the threshold, the processor uses the actual aircraft altitude.
The drawings described herein are for illustrative purposes only of selected embodiments and not all possible implementations. Thus the particular choice of drawings is not intended to limit the scope of the present disclosure.
The following detailed description is merely exemplary in nature and is not intended to limit the invention or the application and uses of the invention. Furthermore, there is no intention to be bound by any theory presented in the preceding background of the invention or the following detailed description.
The disclosed virtual camera system offers a viewing experience that is greatly improved over the conventional video camera system. To put these improvements in context, an understanding of the conventional video camera system is helpful. Therefore, refer to
In the conventional video camera system one or more video cameras are mounted at predefined locations on the aircraft 10, so that such cameras have a view of the ground or airspace, usually from a fixed vantage point. Such cameras are position to give passengers either a view of the runway during takeoff or landing, or a bird's-eye view of the ground beneath the aircraft during flight.
The physical nature of the connection between airframe and conventional camera is shown in
The disclosed virtual camera system is shown in
In its default operating state, the virtual image processor generates an image for presentation on display 36 based on the aircraft's current location in three-dimensional space (e.g., latitude-longitude-altitude). To supply the virtual image processor with the current aircraft location, suitable data connections are provided to the aircraft's location sensor system, which will typically include GPS receiver 42 and altimeter 44.
The aforementioned components of the virtual camera system, the virtual reality image generator system 33, virtual image processor 34, display, georeferenced image data store 38, and the aircraft location sensor system 40, are physically carried by the airframe 10.
The georeferenced image data store 38 is preferably prepopulated (e.g., uploaded or installed prior to flight) with model-based image data files produced by a model-based image system 46, which generates the model-based image data using images of actual Earth terrain captured by a suitable camera/LiDAR system 48. Notably, the model-based image system 46 and camera/LiDAR 48 do not need to be carried by the airframe 10. Indeed, in the illustrated embodiment, the camera/LiDAR 48 and model-based image system 46 are physically and temporally detached from the airframe. In this context, temporal detachment refers to the fact that the image capture by camera/LiDAR 48 and processing by the model-based image system 46 all occur prior to the aircraft flight. Indeed, the image capture and processing need not bear any direct relationship to the aircraft flight at all. All that is required by the virtual image processor 34 are image data suitable for generating a high resolution images for display that are referenced to particular latitude-longitude locations on the Earth.
In one embodiment, the georeferenced image data are constructed using wireframe or mesh models of the surface of the Earth and/or other structures, which define plane surfaces to which a high resolution photographic image is associated. Together the collection of these plane surfaces comprise a photorealistic image of a particular location on the Earth's surface, a waypoint or point of interest or other structure or physical feature. Any of a number of different photogrammetry techniques can be used. Each of the plane surfaces, or at least a cluster of plane surfaces, has associated with it a geotag, identifying the corresponding location on the Earth's surface or structure that portion of the composite image.
One major advantage of using wireframe mesh models is that images can be scaled larger or smaller by simply proportionally increasing the dimensions of the polygonal surfaces (triangles) while retaining an appropriately high resolution of the applied texture images applied to the surfaces.
The output of the model-based image system 46 can be expressed in a variety of different formats, including point clouds, 3D building models, digital elevation models, geospatially corrected aerial images, planemetric feature models (road edges, building footprints, etc.) topographic, terrain and contour maps and volumetric survey data.
In one embodiment, the virtual image processor is designed to provide image content for in-flight display that exceeds the natural viewing resolution of the naked eye. Specifically, the processor can provide a display while the aircraft is at high altitude (e.g., 40,000 feet) that appears as if the aircraft were at 5,000 feet. Alternatively, the processor can provide an out-of-body view of the aircraft in its environment, as if seen from above, below or aside the plane. Such out-of-body perspective may be angled midway between horizontal and vertical, showing a side view of the plane and also features on the ground that are visible from that vantage point.
With reference to
To generate this view, the processor 34 is programmed to first determine the latitude and longitude of the aircraft, as at 54. The latitude and longitude provide the virtual image processor 34 (
As discussed previously, with a conventional aircraft-mounted camera, the viewing altitude is always the actual altitude of the aircraft—and this is so because the camera's lens and sensor 22 are physically attached to the aircraft. However, in the disclosed virtual camera system, the altitude can be selected artificially by the processor at whatever virtual viewing altitude the user requires. Selection of the virtual viewing altitude can be made directly by the user, through manipulation of controls associated with the display 36, or the processor can automatically select the virtual viewing altitude using a preprogrammed algorithm, which can be optionally overridden by the passenger or pilot.
In one embodiment, the processor selects a virtual viewing altitude to match the actual aircraft altitude so long as the aircraft is flying below a predetermined threshold. If this threshold is set at 5000 feet (for example), then the processor will use the actual aircraft altitude as the virtual altitude up to 5000 feet. At altitudes above the 5000 foot threshold the processor will continue to use a virtual altitude of 5000 feet. Thus, when flying at 40,000 feet, the processor will generate a displayed image of the ground as seen from 5000. Of course, the altitude threshold can be set at a different value if required. In this embodiment the virtual altitude is related to the actual altitude through a discontinuous function, such as unit step function:
In another embodiment, rather than use a unit step function, the processor applies a different form of discontinuous function which applies graph-compression to altitudes above the threshold. This can be done, for example, by multiplying altitudes above the threshold by a constant between 0 and 1 and adding this product to the threshold. Such processing will scale the displayed image so that at aircraft altitudes below the threshold (e.g., below 5000 feet) will be accurately represented and at altitudes above the threshold will be compressed.
The resolution of the aircraft display system will influence the database resolution requirements. The ability to host the database may drive the maximum resolution that the aircraft is capable of displaying, assuming the aircraft has some limit on the amount of on-board storage available. For aircrafts that are capable of flying anywhere in the world, the system generally will need to host a global version of the virtual world.
For takeoff, landing, and the ‘virtual low level’ views where the virtual camera is close to the ground, the demand on the graphics processing unit (GPU) of the virtual image processor will be the highest as the scene will potentially be rapidly changing and the rendering will need to be correspondingly faster. This will need to be taken into consideration in determining the technical requirements for the GPU. This may be balanced, particularly for takeoff and landing, by restricting the allowable virtual perspectives because for many people, looking straight down during takeoff and landing can be disorienting and something to be avoided. Restricting this perspective would ease the demand on the GPU without depriving users of the more interesting views.
As demonstrated from the foregoing, the disclosed virtual camera system affords aircraft passengers with an engaging, high resolution view of the terrain during flight. Such view can be augmented, if desired, by including a realistic view of the aircraft. Under normal operation, the view provided by the virtual camera system is associated with the actual location of the aircraft relative to the ground below; however, the apparent altitude of the virtual image may be adjusted to give the viewer a much closer view of the ground, as if the aircraft altitude were much lower. Thus the virtual image is made to appear magnified, as if viewing through a telephoto lens, but without any optical distortion. Such magnification is made possible because the virtual image is digitally scaled by the virtual image processor by scaling the size of the triangles forming the triangulated irregular network.
Compared with images from conventional video cameras, the images produced by the virtual camera system are strikingly realistic. There is no blurring or optical distortion which often accompanies conventional camera systems which are necessarily exposed to the weather. Unlike conventional camera systems, the virtual camera system provides views of the ground that are not reduced to microscopic size due to high altitude, nor occluded by clouds. These benefits result from the fact that the camera/LiDAR system 48 (
Such viewing experience is not a real-time viewing experience, because the images were captured in the past. However, giving up the real-time experience, passengers are rewarded with a much more visually stimulating, high resolution view of the Earth and structures erected thereon. Moreover, the passengers can manipulate the perspective of their viewing experience using convenient user-interface controls associated with the display 36 (
While at least one exemplary embodiment has been presented in the foregoing detailed description, it should be appreciated that a vast number of variations exist. It should also be appreciated that the exemplary embodiment or exemplary embodiments are only examples, and are not intended to limit the scope, applicability, or configuration of the invention in any way. Rather, the foregoing detailed description will provide those skilled in the art with a convenient road map for implementing an exemplary embodiment as contemplated herein. It should be understood that various changes may be made in the function and arrangement of elements described in an exemplary embodiment without departing from the scope of the invention as set forth in the appended claims.