Gas turbine engines are required to be inspected, usually at predetermined intervals according to a maintenance schedule. Such inspections are often invasive, time-consuming, and expensive because the engine (and thus aircraft) is out-of-service during inspection. Inspections, however, are key to identifying distress in engine components, which may then require repair or replacement. Some inspections need to be performed after each flight and can be difficult to conduct.
A system for inspection of a gas turbine engine according to an example of the present disclosure includes at least one drone operable for flight and equipped with an imaging device and a light source, and at least one processor configured to operate the at least one drone to fly into a gas turbine engine to a first position with respect to a component in the gas turbine engine, operate the imaging device and the light source to take an image a target surface of the component from the first position, and identify whether the image includes an obstruction blocking a portion of the target surface from view of the imaging device. In response to identifying the obstruction, the at least one drone operates to fly to a second position from which there is a line-of-sight to the target surface without the obstruction, and operates the imaging device and the light source to obtain an unobstructed image of the target surface from the second position.
In a further embodiment of any of the foregoing embodiments, the component is a fan blade and the obstruction is an inlet guide vane.
In a further embodiment of any of the foregoing embodiments, the processor includes one or more neural networks configured to identify whether the image includes the obstruction.
In a further embodiment of any of the foregoing embodiments, the one or more neural networks is configured to identify an abnormality in the target surface from the image.
In a further embodiment of any of the foregoing embodiments, the one or more neural networks is configured to navigate the at least one drone.
In a further embodiment of any of the foregoing embodiments, the at least one drone includes first and second drones. The imaging device of the first drone takes the image of the target surface from the first position and the image device of the second drone taking the unobstructed image from the second position.
In a further embodiment of any of the foregoing embodiments, the imaging device includes a borescope.
A further embodiment of any of the foregoing embodiments includes a docking station on an aircraft associated with the gas turbine engine from which the at least one drone is deployed to fly into the gas turbine engine.
A further embodiment of any of the foregoing embodiments includes an operator interface configured to permit an operator to take images using the imaging device.
A method for inspection of a gas turbine engine according to an example of the present disclosure includes operating at least one drone to fly into a gas turbine engine to a first position with respect to a component in the gas turbine engine. The at least one drone is equipped with an imaging device and a light source. The imaging device and the light source are operated to take an image a target surface of the component from the first position and identify whether the image includes an obstruction that blocks a portion of the target surface from view of the imaging device. In response to identifying the obstruction, the at least one drone operates to fly to a second position from which there is a line-of-sight to the target surface without the obstruction, and operates the imaging device and the light source to obtain an unobstructed image of the target surface from the second position.
In a further embodiment of any of the foregoing embodiments, the identifying of whether the image includes an obstruction is performed using one or more neural networks.
A further embodiment of any of the foregoing embodiments includes identifying from the image whether the target surface includes an abnormality.
In a further embodiment of any of the foregoing embodiments, the at least one drone includes first and second drones, and includes coordinating operation of the first and second drones to take the image of the target surface from the first position with the imaging device of the first drone and take the unobstructed image from the second position with the imaging device of the second drone.
In a further embodiment of any of the foregoing embodiments, the imaging device includes a borescope, and including operating the drone to deploy the borescope to take the unobstructed image.
A further embodiment of any of the foregoing embodiments includes operating the at least one drone to deploy from a docking station on an aircraft that is associated with the gas turbine engine.
A further embodiment of any of the foregoing embodiments includes manually taking the image through an operator interface that is configured to operate the imaging device.
The present disclosure may include any one or more of the individual features disclosed above and/or below alone or in any combination thereof.
The various features and advantages of the present disclosure will become apparent to those skilled in the art from the following detailed description. The drawings that accompany the detailed description can be briefly described as follows.
In this disclosure, like reference numerals designate like elements where appropriate and reference numerals with the addition of one-hundred or multiples thereof, if used, designate modified elements that are understood to incorporate the same features and benefits of the corresponding elements. Terms such as “first” and “second” used herein are to differentiate that there are two architecturally distinct components or features. Furthermore, the terms “first” and “second” are interchangeable in that a first component or feature could alternatively be termed as the second component or feature, and vice versa.
The system 20 includes at least one drone 26 and a controller 28. The drone 26 may also be referred to as an unmanned aerial vehicle. The controller 28 is in communication with the drone 26, such as by a wired or wireless connection 30. As an example, the drone 26 is a propeller-based, multirotor design, such as a quadcopter. In general, the drone 26 is miniature in size so as to be able to enter into the engine 22 and maneuver therein with clearance to be able to rotate and traverse the engine components being inspected.
In the example shown, the system 20 includes a docking station 32 from which the drone 26 deploys for inspection of the engine 22. For example, the system 20 as shown is a “non-dedicated” system that is remotely located from the aircraft 24. In that regard, the docking station 32 may be mobile, such as on a wheeled cart or vehicle or small enough to fit in a hand-carried case or backpack such that it can be moved to into proximity of the engine 22 or moved to service multiple engines across different aircrafts. Alternatively, however, the system 20 is a “dedicated” system that is intended to service the engine or engines of a single aircraft and, in that regard, the docking station 32 is on the aircraft 24, as designated at 132. The docking station 32 (or 132) serves as a point for take-off and landing of the drone 26, but may also be configured to refuel or recharge the drone 26, and may include a computerized device or other electronics for communicating with and/or controlling flight and operation of the drone 26.
The processor 34 of the controller 28 is configured to operate the drone 26, in accordance with instructions stored in the memory 36, to fly into the gas turbine engine 22 for inspection and perform the functions described herein. The following examples are based on inspection of fan blades 48 of the engine 22, although it is to be understood that the system 20 is not limited and will be applicable for inspection of other engine components, as well as non-engine implementations. The processor 34 operates the drone 26 to fly to a first position P1 with respect to the fan blades 48, as shown in
Once at the first position P1, the processor operates the imaging device 40 and light source 42 of the drone 26 to take an image 50 of a target surface 48a of the fan blades 48. In this case, however, there is an obstruction 52 (shown in phantom) between the imaging device 40 and the target surface 48a. In this example, the obstruction 52 is an inlet guide vane located forward of the fan blades 48 that blocks the view of a portion of the target surface 48a that is to be inspected. Thus, full inspection cannot be completed based on the image 50 from the first position P1. The obstruction 52 is not limited to guide vanes and other examples include, but are not limited to, platforms, casings, and adjacent blades that overlap so as to be in the field of view.
Of course, not all images will necessarily be obstructed, and the processor 34 thus performs an analysis to identify whether the image 50 includes an obstruction. In this regard, the processor 34 (
If no obstruction is identified in the image 50, the processor 34 operates the drone 26 to move on to a new first position for inspection of another target surface, such as another portion of a blade or blades 48. If, however, an obstruction is identified, the processor 34 responsively operates the drone 26 to fly to a second position P2, as also shown in
In further examples, the positions P1 and P2 may take into account information that the drone 26 has about the engine 22. For instance, the drone 26 may have information about the engine model, number of blades 48, the size and spacing of the blades 48, and the presence, size, and spacing of guide vanes stored in the memory 36 such that the drone 26 can move and operate more intelligently with respect to the locations of yet-to-be inspected blades 48 and already-inspected blades 48. For instance, the size and spacing of the guide vanes may not permit any single position, or even a combination of multiple positions, from which a fully unobstructed image or images can be taken. In that case, knowing beforehand that reiteratively moving to new second positions P2 would be in vain, the drone 26 may instead move to a second position in which the borescope 44 can be deployed past the guide vanes to get a clear image of a target surface.
In a further example in which the drone 26 does not have information of the number of blades 48, the drone 26 may utilize the imaging device 40 to identify and “count” the number of blades 48 as it inspects or prior to beginning inspection, such that it can determine whether all of the blades have been inspected for completion of an inspection mission.
In a further example, rather than a single drone 26, there are two or more drones 26 that operate in coordination. For instance, rather then the single drone 26 moving from the first position P1 to the second position P2, the processor 34 coordinates operation of a second drone 26 to move to position P2 to take the image 50a.
The processor 34 is also operable to identify whether the target surface 48a contains an abnormality 56 (
Although a combination of features is shown in the illustrated examples, not all of them need to be combined to realize the benefits of various embodiments of this disclosure. In other words, a system designed according to an embodiment of this disclosure will not necessarily include all of the features shown in any one of the Figures or all of the portions schematically shown in the Figures. Moreover, selected features of one example embodiment may be combined with selected features of other example embodiments.
The preceding description is exemplary rather than limiting in nature. Variations and modifications to the disclosed examples may become apparent to those skilled in the art that do not necessarily depart from this disclosure. The scope of legal protection given to this disclosure can only be determined by studying the following claims.