A Health-Monitoring process, commonly known as PHM (for Prognostics & Health Management) although it is also referred to, but less frequently, as condition-based maintenance, predictive maintenance or forecast maintenance, is used to anticipate the degradation of an equipment or a system, particularly in the aeronautical field.
A Health-Monitoring process thus makes it possible to avoid breakdowns, optimise service life, anticipate and plan maintenance operations (and possible removal), and thus reduce the costs of equipment or system maintenance operations.
A Health-Monitoring process consists of five main steps: data acquisition, data processing (or preconditioning), signature extraction, diagnosis and prognosis.
The data-acquisition step consists of acquiring data produced by sensors measuring parameters representative of a performance of the equipment or the system. Data acquisition is for example carried out via testing, or by downloading to the equipment or system when it is operational.
The data-preconditioning step includes filtering data to remove measurement noise from the data.
The signature-extraction step consists of processing the data to extract one or more signatures representative of the state of the equipment or system and, possibly, of a degradation of the operation of the equipment or system.
The diagnosis step consists of determining the state of health of the equipment or system at the current time from a set of fault classes (failure modes), each associated with a severity level.
The prognosis step consists in estimating the Remaining Useful Life)(RUL) of the equipment or system.
The purpose of the invention is to reliably and accurately estimate the Remaining Useful Life of a piece of subject equipment.
In order to achieve this goal, a method for estimating a Remaining Useful Life of a subject equipment is proposed, with a preliminary phase comprising the following steps:
The estimation method according to the invention therefore implements, during a preliminary phase carried out for example in the laboratory, a first learning of a diagnosis model and a second learning of a prediction model on at least one signature resulting from observations made on test equipment devices.
Then, during an operational phase, observations are made in real time on the subject equipment, and, at a “present” time t, the evolution of the signature on the subject equipment is extrapolated thanks to the prediction model and then classified thanks to the diagnosis model. The severity class membership function then provides a reliable and accurate estimate of the Remaining Useful Life of the subject equipment, which is based on the behaviour of test equipment devices under similar stresses, the test equipment devices being equipment similar to the subject equipment.
An estimation method such as the one just described is also proposed, in which the estimated Remaining Useful Life is such that:
=q−t
where q is an estimated end-of-life time and t is a current time, and where, at the estimated end-of-life time q, the last severity class membership function becomes greater than the penultimate severity class membership function.
An estimation method such as the one just described is also proposed, in which a temporal consistency criterion is used to define a number of severity classes produced by the partitioning.
In addition, an estimation method such as the one just described is proposed, in which the temporal consistency criterion is evaluated using a first metric defined by:
where xk(i) is an ith test observation acquired for test equipment k and where ct is a temporal consistency between two successive test observations and is defined using the following formula:
where yk(i) is the severity class of the observation xk(1).
An estimation method such as the one just described is also proposed, in which a second metric, which is an energy metric, is evaluated to select a partitioning algorithm.
An estimation method as described above is further proposed, where the energy metric is defined by:
where Ωi is a set of test observations of an ith class, the x(j)∈p are the test observations, gi is a centre of gravity of the ith severity class, d is a distance measure and is defined by:
d
ij
2
=d(x(j),gi)2=∥x(j)−gi∥d2=(x(j)−gi)TAd(x(j)−gi)
where Ad is a positive semidefinite matrix defining a distance metric, and where the symbol “∥ ∥d” represents the norm corresponding to the distance d.
An estimation method such as the one just described is also proposed, in which the partitioning algorithm is a Fuzzy C-Mean algorithm.
An estimation method such as the one just described is also proposed, in which the prediction model is trained with a multi-step iterative method.
An estimation method such as the one just described is also proposed, in which the prediction model is constructed from a non-linear SVR regression.
In addition, an estimation method such as the one just described is proposed, in which the first learning of the diagnosis model uses a Fuzzy SVM-FA classification algorithm.
The invention will be better understood when reading the following description of a particular non-restrictive embodiment of the invention.
Reference is made to the appended drawings, wherein:
The present invention is implemented in a Health-Monitoring process, and relates to a method for estimating the remaining useful life of a subject equipment.
The invention therefore relates more specifically to the “prognosis” stage of the Health-Monitoring process.
Here, by “prognosis” and according to the ISO 13381 standard, we mean: “the estimation of the duration of operation before failure and the risk of the existence or subsequent occurrence of one or more failure modes”. By adopting a system-oriented interpretation that takes into account the interactions between the components of the system, the component is considered to reach its end of life when its state of damage causes or accelerates the damage of other components. The term “subject equipment” should be interpreted very broadly. The subject equipment may be any equipment, for example an LRU (Line Replaceable Unit) such as an actuator, or a system comprising several pieces of equipment interacting with each other, or an electrical or mechanical component of any complexity. The subject equipment is here embodied in an aircraft, but the invention is not limited to such an application.
It is assumed in the following that the subject equipment is not being rejuvenated and is not being subjected to any maintenance action that would restore it to a previous state of health. Thus, its Remaining Useful Life (RUL) is linearly decreasing with time t:
RUL=tEOL−t,
where tEOL is the End of Life of the subject equipment.
Time can be considered in different ways depending on the application.
The time since commissioning can be counted. The cumulative running time can also be used. For an electromechanical actuator, this corresponds to the phases when the electromechanical actuator is energised. Here, in the context of an aeronautical system, which is very strict in terms of component traceability, both time information is available. The time taken into account could also be the number of cycles of use when it comes to the considered deterministic speed/load profiles.
Here, the cumulative operating time is used.
It is also considered that the subject equipment degrades only during operation. Indeed, for an actuator of an aeronautical system, and in particular for an actuator integrated in a single-aisle aircraft, the waiting phases, on the tarmac or in the hangar, are low compared to the operating times.
The estimation process comprises a plurality of steps. These steps are implemented in a preliminary phase and then in an operational phase. The preliminary phase is carried out on test equipment devices similar to the subject equipment, and is at least partially implemented in a laboratory or design office. The operational phase is implemented while the subject equipment is in operation.
Each step will be described theoretically and then illustrated with an application example.
With reference to
This public data is provided by NASA (Lee J, Qiu H, Yu G, Lin J, Rexnord Technical Services. Bearing Data Set [Internet]. IMS, University of Cincinnati, NASA Ames Prognostics Data Repository (http://ti.arc.nasa.gov/project/prognostic-data-repository), NASA Ames Research Center, Moffett Field, Calif.; 2007. They are available on: https://ti.arc.nasa.gov/tech/dash/pcoe/prognostic-data-repository/Qiu et al. 2006; Lee et al. 2007). Four roller bearings 1a, 1b, 1c, 1d are placed on a single shaft 2 driven by a motor 3 to rotate at 2000 rpm. A spring mechanism applies a constant radial load of 6000 pounds, or 26,698 N.
With reference to
We carry out mu observations on nu test equipment devices.
In the application example, the observations are made by means of accelerometers 4 integrated in the bearings 1.
The preliminary phase of the estimation process then includes a step of preconditioning the data 11.
The preliminary phase then includes a signature extraction step 12.
The vector of fault signatures is denoted by X: X=[x1, x2, . . . , xp], where p is the number of fault signatures. A signature is a quantity that is sensitive to the degradation of the test equipment devices (and the subject equipment).
The signatures that are used here are the power of the vibration signal and the RMS value of the vibration signal.
The signatures are stored in a database 13.
Time series signatures are then produced. Each SX signature time series corresponds to the evolution of the signature over time.
The estimation process then includes an automatic partitioning step 14 of the data to create successive classes. The classes are also called failure modes, failure levels or severity classes. The classes correspond to ageing phases of the test equipment devices.
The goal of the partitioning is to identify, from the mu observations made on the nu test equipment devices, the different severities of defects present in the database. Thus, it will be necessary to partition m=mu*nu observations. Each observation corresponds to a vector of shapes with p parameters. A test equipment device is therefore represented by p time series (one time series per shape vector signature). The resulting severity classes should be temporally consistent across all test equipment devices.
For each test equipment device, each class will be defined by a start time tstart and an end time tstop. The severity classes are arranged from least to most severe. Thus, the tstop of a class will be the tstart of the next class.
In the example in
First of all, we try to define an optimised number of classes.
The partitioning of time series classically contains three main categories of problems: partitioning of whole time series clustering, subsequences clustering and time point clustering.
The partitioning of whole time series consists of considering each series as an object and forming groups of whole series. This type of partitioning is equivalent to classical partitioning but using integer time series rather than shape vectors. The aim of subsequence partitioning is to determine subsequences that repeat in a single time series. The partitioning of time series points consists of assigning classes based on temporal proximity and values within a single series.
None of these categories corresponds exactly to the problematic of the invention. Partitioning entire time series is not suitable, since the objective is to find successive sequences. Sub-sequence partitioning is also not suitable as it works from a single sequence. The same applies to the partitioning of time points. The problem here is to deal with nu*p sequences, linked to each other.
A criterion of temporal consistency of the partitioning is therefore defined and used to choose the number of classes.
A definition of temporal consistency of partitioning is given here. It is based on three conditions:
The following first metric is proposed to evaluate the temporal consistency on observations of a test equipment device, which are ordered in increasing time:
where xk(i) is the ith observation measured for test equipment devices k and where ct is the temporal consistency between two successive observations and is defined using the following formula:
where yk(i) is the class of the observation xk(i).
For a given test equipment devices k, the resulting partitioning is perfectly consistent for CTk=c−1, where c is the number of classes.
Conversely, if the partitioning is totally inconsistent, we have CT=mu−1 where mu is the number of observations. To facilitate the interpretation of the first metric and its use for comparisons, an equivalent version is defined with values in [0,1]:
For a perfectly temporally consistent partitioning, we have:
CTN=0.
For a completely inconsistent partitioning, we have:
CTN=1.
For a partitioning where the number of classes is different from one test equipment devices to another, we have:
CTN<0.
We can now calculate the overall temporal consistency, over all the test equipment devices, using the following formula:
where “|” is the absolute value symbol.
The first metric verifies the first two conditions, i.e. the number of identical classes per test equipment devices and temporal continuity, but not the third condition. The third condition can nevertheless be verified by plotting the evolution of signatures over time for partitioned observations.
A partitioning algorithm should also be selected. For a given number of classes, if several partitioning algorithms are perfectly consistent, a second metric is evaluated to select the partitioning algorithm. The second metric is an energy metric, which must be minimised to have the most compact classes possible. The second metric is calculated using the following formula:
where Ωi is the set of observations of the ith class, the x(j)∈p are the observations, gi is the centre of gravity of the ith class, d is the distance measure and is defined by the following formula:
d
ij
2
=d(x(j),gi)2=∥x(j)−gi∥d2=(x(j)−gi)TAd(x(j)−gi)
where Ad is a positive semidefinite matrix defining the distance metric d (chosen by the user), and where the symbol “∥ ∥d” represents the norm corresponding to the distance d.
Here, the Euclidean distance is adopted with:
A
d
=I, the identity matrix.
Four algorithms, classically used for unsupervised partitioning, were evaluated: the K-means algorithm, the K-means algorithm, the C-means fuzzy algorithm and the C-Fuzzy means possibility algorithm.
With reference to
We notice that for c>6, all partitionings are inconsistent, whatever the chosen algorithm. We then notice that only the Fuzzy C-Mean algorithm gives consistent results between c=4 and c=6.
With reference to
The curve corresponding to the C-Mean Fuzzy Possibilities algorithm is off the graph, as this curve reaches too high values. It is clear that the Fuzzy C-Mean algorithm has the lowest energy, whatever the number of classes on the interval considered. This algorithm is therefore selected.
Note that the C-Fuzzy means algorithm was developed as an improvement of the K-Fuzzy algorithm. It is considered that an observation can belong to several classes at the same time but with different degrees. With reference to
With reference to
The signature of interest in
The signature of interest in
It can be see that, for this number of classes, the temporal consistency is preserved and the level of convergence of the energy criterion has reached a reasonable level (reduced by two thirds). The first class corresponds to the run-in phase. The end-of-life phase is well insulated. The living environment is divided between two classes.
With reference to
The preliminary phase of the estimation process according to the invention then comprises a diagnosis step 44 (see
The diagnosis consists of assigning each operating point to its nearest class for a given distance type. A classification model must therefore be trained to perform this task using data available in the dedicated database, said data coming from the test equipment devices.
It would have been possible to use the classification method shown in
In such a method, a similarity between a new observation and classes c1, c2, c3 is measured, with class c1 corresponding to no defect, class c2 corresponding to a first defect and class c3 corresponding to a second defect.
This classification is discrete and does not take into account the level of membership in a given class. However, another level of classification is chosen to assess the severity level of each failure mode, using fuzzy membership functions.
The first learning of the severity diagnosis model is based on the Fuzzy SVM (Space Vector Machine) classification algorithm with multi-class membership functions.
The diagnosis model is then implemented (step 46) and validated (step 47). The result of the classification with severity level can be seen in
The preliminary phase of the estimation process according to the invention then comprises a prognosis step 50 (see
The prognosis consists of extrapolating the signature into a finite horizon, which we will call the prediction horizon (step 54 in
First, the construction phase of the prediction model (which can also be called a regression model) is described.
A learning method for the prediction model is chosen. Predictions are made at several time steps, based on so-called “multi-step” strategies.
These predictions are based on a sequence of observations of past moments. This expresses itself under the form:
{circumflex over (X)}
t+1→t+hp
=f
p(Xt−n0→t,θ)
where:
The number of observations and the prediction horizon are usually set by the user.
If all the coefficients of {circumflex over (X)}t+1→t+hp∈, we speak of uni-variate extrapolation. If the coefficients of {circumflex over (X)}t+1→t+hp∈p, we speak of multivariate extrapolation. This second case allows correlations between signatures to be taken into account.
The multi-step iterative method consists of learning a model to make predictions at t+1. Then the same model is iteratively run in series to make a prediction at t+hp.
The advantage of this method is that it is fairly easy to implement. The disadvantage is that error accumulates as predictions are made. Within the framework of aeronautical constraints, it is necessary to take into account a dimensional factor: the memory required to store the models. In view of these constraints, the iterative method seems to be a good choice as a first approach.
The algorithm for this method is summarised in
The linear regression model itself is now described.
The purpose of the time regression is to determine the parameter vector θ that determines the fp prediction model. This model, which can also be called a learning model, is built from a non-linear SVR (Support Vector Machine for Regression) regression. The objective of the latter is to provide a model predicting an output y∈ from a set of m labelled data {(x1, y1), . . . , (xm, ym)}ou xi∈p, p∈*, i∈{1, . . . , m}
The algorithm must therefore find a function fp: p→, which minimises the error on the learning set while having a good generalisation capacity. The regression problem can be formulated using the support vector machine formalism. This is based on the SVM with flexible margins. For a non-linear case, we use the kernel trick to reformulate the problem. The aim is to maximise the following function:
Under the constraints:
where K is the kernel function which can be chosen such that:
K(x,x′)=tanh(ak·x·x′+bk),
and where (αi, αi*) are the primary and dual Lagrange coefficients. The prediction function is then expressed as:
The constant b can be obtained from the KKT (Karush-Kuhn-Tucker) conditions applied on one of the support points, i.e. the points corresponding to:
for:
The prediction models are then implemented on all signatures (step 52) and validated (step 53).
The above description is applied to the roller or ball bearings 1 shown in
SVR is implemented via the Statistics and Machine Learning Toolbox™ in Matlab®. Three prediction start times were tested: a start time at t=sp1=1 day (where no observations of the test equipment devices are used for learning), a start time at t=sp2=10.45 days, which corresponds to the end of the run-in phase, and a start time at t=sp3=20.65 days, which corresponds to the start of the penultimate severity class.
The prediction results are evaluated according to the Mean Absolute Percentage Error: MAPE), which is the dual of the cumulative relative accuracy.
The prediction results are equivalent for test equipment devices.
Curve 60 corresponds to the start at t=sp1, curve 61 to the start at t=sp2, and curve 62 to the start at t=sp3. Curve 63 corresponds to the target, i.e. the actual development observed.
Curve 64 corresponds to the start at t=sp1, curve 65 to the start at t=sp2, and curve 66 to the start at t=sp3.
Curve 67 corresponds to the start at t=sp1, curve 68 to the start at t=sp2, and curve 69 to the start at t=sp3
The error increases with time, which is consistent with the use of an iterative method. In general, the greater the number of observations used for the subject equipment, the smaller the error. The error increases sharply for the last predictions. This is due to an overestimation of the exponential growth of the signature. However, this is after the end-of-life time of this unit, which is 28 days.
With reference to
The MDLC classification models as well as the MDLR prediction models are considered to be already trained.
The operational phase includes the step of acquiring observations when in operation on the subject equipment. Signature sequences are produced.
Extrapolated time series representative of the evolution of signatures on the subject equipment are produced using the prediction model.
Predicted signatures from t+1 to t+hp are obtained.
The extrapolated time series are classified using the diagnosis model.
Membership functions are calculated for the time series extrapolated to the severity classes. From these membership functions, the Remaining Useful Life of the subject equipment is derived.
The membership functions uj (j going from 1 to c) indicate the assigned failure class and its severity level.
The prognostic horizon is the gap between the time tαeβe when the algorithm reaches the performance αe (accuracy) and βe (precision) and the end-of-life time tEOL is written as follows:
PH=tEOL−t+eβe.
With reference to
The objective here is only to provide a trend, well in advance, so that maintenance operations can be anticipated.
Thus, the constant αe will be defined to have an interval of ±15% around the initial RUL.
The implementation of the relative precision does not require any additional information. In contrast, the prognosis horizon and performance require further information.
Since the RUL is not given here as a distribution, the confidence level βe does not have to be defined. The prognosis horizon will therefore be calculated by:
PH=tEOL−tαe, where:
tαe is the time from which the RUL estimate checks until the end of life:
RÜL∈[(1−αe)*RUL(0);(1+αe)*RUL(0)]
The best horizon is obtained when the algorithm is always in the desired performance range and the worst when it is never there. In this case, αe is the uncertainty interval on the initial RUL, RUL(0).
The RUL(Remaining Useful Life) corresponds to the remaining life time from the current time t. The classification model then calculates the memberships uj (j ranging from 1 to c) of the severity classes. From this, the RUL is deduced as follows.
The estimated RUL is equal to the difference between the current (prediction) time minus the estimated end-of-life time q. Therefore, with 1≤q≤hp:
=q−t
where q is an estimated end-of-life time and t is a current time, and where, at the estimated end-of-life time q, the last severity class membership function becomes greater than the penultimate severity class membership function.
The membership of the last class c becomes higher than that of class c−1, and thus:
u
c−1({circumflex over (X)}t+q)<uc({circumflex over (X)}t+q)
As an example, the graph in
Curve 70 corresponds to the u1 class 1 membership function, curve 71 corresponds to the u2 class 2 membership function, curve 72 corresponds to the u3 class 3 membership function and curve 73 corresponds to the u4 class 4 membership function.
The limit 74 corresponds to the beginning of class 4, i.e. the last class, and therefore to the end-of-life time.
Thus, the end-of-life time is estimated to be:
q=27.23 days and the remaining lifetime:
=q−t=27.23−10.45=16.78 days.
In addition to the prediction t=sp2=10.45 days, the estimation of the Remaining Useful Life of the 1d roll was also carried out by two other prediction models t=sp1=7.5 days and t=sp3=20.65 days.
Curve 75 is the actual value of the remaining useful life. Curve 76 corresponds to the lower tolerance and curve 77 to the upper tolerance. Curve 78 corresponds to the prediction t=sp1, curve 79 corresponds to the prediction t=sp2, curve 80 corresponds to the prediction t=sp3.
It can be seen that all three estimates are very close to the actual RUL, which shows that the estimation process according to the invention is very accurate. Of course, the invention is not limited to the described embodiment but encompasses any alternative solution within the scope of the invention as defined in the claims.
Number | Date | Country | Kind |
---|---|---|---|
RE1907561 | Jul 2019 | FR | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/EP2020/068705 | 7/2/2020 | WO |