This Application claims priority of China Patent Application No. 201711071217.1, filed on Nov. 3, 2017, the entirety of which is incorporated by reference herein.
The invention belongs to air traffic operation and management in general. More particularly to an airspace situation evaluation framework based on machine learning and knowledge transfer.
Worldwide, the air transport industry is growing rapidly, accelerating the movement of goods and people but creating even more challenges to the Air Traffic Management (ATM) system. Especially, large air transport volume imposes high workload on air traffic controllers (ATCos), which is the primary cause of operational errors. For instance, a study driven by the National Aeronautics and Space Administration in 2010 found that ATCos suffer from “chronic fatigue” that has resulted in flight navigation errors, like allowing airplanes to fly too closely together. Besides, in 2014, two on-duty ATCos at the control tower of Wuhan airport, China, fell asleep due to high work pressure, and an approaching flight has to go around in absence of communication with the control tower. Thus, how to evaluate ATCo's workload holds great significance for air traffic safety.
In the current ATM system, a sector is the fundamental airspace unit for traffic operation and control service. During the flying process from departure to destination airport, a flight will successively pass through several sectors with the guidance of ATCos. For properly adjusting the control workload of each sector, it is important to accurately evaluate the sector situation (SS), which actually determines the difficulty of traffic control and naturally an indicator to evaluate whether the ATCo is over-burdened or not. Moreover, SS evaluation can also be applied as a guidance of traffic management, such as rearranging sector configurations and traffic flow, thus it becomes a prevalent research direction in the ATM domain.
Due to numerous interacting factors, people resort to machine learning models for SS evaluation. The machine learning model is capable of constructing the complicated mappings between SS and its correlated factors. The current related methods of machine-learning-based SS evaluation can achieve satisfactory evaluation performance as long as the sample set is sufficiently large. However, due to the following two reasons, we have to evaluate SS with a small sample set: (i) The participation of ATM experts is essential in the sample labeling process. Hence, the sample collection is a time-consuming labor-intensive work that we cannot afford easily; (ii) A machine learning model generally requires that the training samples and the samples to be classified be consistent in every aspect, i.e., the distribution of factors' values and the rules of sample categorization. However, the situation producing mechanisms of sectors can be quite different. Hence, normal machine learning models cannot directly utilize the samples of other non-target sectors for the SS evaluation task of the target sector (For simplicity, we use “target sector” to refer to the sector for which the SS evaluation is carried out, “non-target sectors” for any other sectors apart from the target sector, and “target/non-target samples” for samples of target/non-target sectors). This can further aggravate the lack of training samples. Therefore, it is necessary to develop a SS evaluation model that is capable of sufficiently learning on both target and non-target samples to obtain the ability of accurately evaluating the target samples' SS.
The objective of the present invention is to provide a sector situation (SS) evaluation framework that is capable of sufficiently learning on both target and non-target samples, aiming to accurately evaluate the target sector's SS. The SS evaluation framework is based on knowledge transfer and is specifically applicable for small-training-sample environment. The SS evaluation framework is able to effectively mine knowledge hidden within the samples of both target and non-target sectors, and properly handle the integration between the SS evaluation knowledge derived from different sectors.
This SS evaluation framework includes a server and a memory module storing airspace operation data and connected to the server via a local area or wide area network. The airspace operation data includes aircraft trajectory data and airspace configuration data, etc. The SS evaluation framework evaluates the airspace situation by executing steps including:
(1) sufficiently mine the knowledge within the samples of the target sector using the strategies of multi-factor subset generation and multi-base evaluator construction, and multiple target base evaluators will be built after executing these strategies;
(2) precisely learn the knowledge in the samples of the non-target sectors using similar strategies (multi-factor subset generation and multi-base evaluator construction) for the target samples, together with a sample transformation strategy, and multiple non-target base evaluators will be built after implementing these strategies: and
(3) efficiently integrate the target and non-target base evaluators based on evaluation confidence analysis of those target and non-target base evaluators.
Embodiments will now be described by way of non-limiting examples in conjunction with the following figures:
To evaluate the situation of a target sector, two fundamental tasks have to be done first: (i) identify the airspace situation factors as comprehensive as possible (subsequently, the single word “factor” is used to refer to the airspace situation factor). These factors, by which SS is evaluated, are the attributes, parameters and ingredients of airspace and traffic that can influence or reflect the airspace situation. To date, many scholars in the ATM field have proposed a great many of factors, whose values are generally calculated based on some raw airspace operation data, like aircraft trajectory data, airspace configuration data, etc. Table 1 exemplifies some factors (wherein similar factors are grouped together); (ii) collect as many airspace situation samples of the target sector as possible (subsequently, the single word “sample” is used to refer to the airspace situation sample). One sample corresponds to the traffic scenario within the target sector of one certain time slice, and it comprises the values of the factors and the SS condition corresponding ta the values of those factors. The SS condition, which is the label of a sample, is provided by ATM experts. It can be characterized as a continuous index, e.g., sector congestion degree, or discrete levels, e.g., low/normal/high complexity.
In this embodiment, it is assumed that the SS level (low/normal/high complexity) of the target sector will be evaluated based on m factors. Therefore, a number of m-dimensional samples will be collected and learned on to construct the correlations between the factors and the situation levels. Besides the samples of the target sector, it is also possible to obtain and save up the samples of other non-target sectors. This is because, in real applications, there usually exist multiple target sectors whose situation needs to be evaluated. In such a case, a target sample set for each one of the target sectors needs to be collected and each target sample set is in effect a non-target sample set with respect to other target sectors. Note that the non-target samples can actually be utilized for training the SSEF of the target sector through a certain means despite the discrepancies between target and non-target samples. Therefore, we propose a SSEF which is able to use both target and non-target samples for the situation evaluation task of the target sector.
In block 401, multiple factor subsets (FSSs) are generated through selecting factors from FP 402. In order to eliminate noisy and redundant factors within each FSS, the FSS needs to be generated under the a priori knowledge of factor criticality and independence. Hence, factor analysis needs to be implemented. The criticality of each factor is calculated based on the target dataset D(tar)={X(tar),Y(tar)} (block 403). And it is measured by the criterion of signal-to-noise ratio (SNR), which is defined as
where SNR(ft) is the SNR of the factor ft. μL(ft) and σL(ft) are, respectively, the mean value and the standard deviation of ft's values attributed to the Category L. Similarly, μN(ft), σN(ft), μH(ft) and σH(ft) are the corresponding statistics of ft's values attributed to the Categories N and H. It is easy to find that the SNR value of a factor reflects the correlation degree between this factor and sample category (SS level). A larger SNR value indicates a stronger ability to discriminate among sample categories, that is, more critical to the situation evaluation task.
Besides measuring the criticality of each factor, the independence coefficient (IC) is used to evaluate the independence degree between each two factors. Its formula is
where ICft2/ft1 measures the independence of the factor ft2 from another factor ft1 (generally, ICft2/ft1≠ICft1/ft2); ft1i and ft2i are respectively the values of ft1 and ft2 within the i th sample (1≤i≤ntar);
ft2(ft1) is the regression function between ft1 and ft2 which is obtained from the regression analysis based on the target samples. Specifically, ft2(ft1) can be obtained through the locally weighted linear regression method. In this method, the expression of the regression function is
ft2(ft1)=[1, ft1](BTWft1B)−1BTWft1ft2,
where ft2=[ft21, ft22, . . . , ft2n
Wft1=diag(wft1)(wft1=[wft1(ft11), wft1(ft12), . . . , wft1(ft1n
Next, T FSSs 404 are generated, each of which contains k non-duplicated factors in FP 402 (k<m). The FSS generation process is implemented under the guidance from the a priori knowledge on the factor's criticality and independence to prevent noisy and redundant factors from being selected into each FSS.
After T FSSs 404 are generated, factor reduction is implemented on X(tar) according to each FSS, so that finally T dimension-reduced target sample subsets 405 are obtained (wherein each target sample subset comprises a dimension-reduced X(tar) and Y(tar)) Based on each target sample subset, a target base evaluator (among the target base evaluators 406) is built (subsequently, “target/non-target base evaluator” is used to refer to the base evaluators trained by target/non-target samples). Thus, T target base evaluators 406 are finally obtained.
In
Here we elaborate on the procedures of the non-target sample subset transformation in block 607. Let D(ntdr)={X(ntdr),Y(nt)} denote a non-target sample subset (among the T non-target sample subsets 606), where X(ntdr)=[x1(ntdr), x2(ntdr), . . . , xn
The first sample transformation step is mapping X into a Reproducing Kernel Hilbert Space with a kernel g. Let g(x,⋅) denote the element in to which the sample x is mapped. Generally, g(x,⋅) is a transformed sample of infinite-dimensionality. Thereupon, X is mapped to g(X,⋅)=[g(x1,⋅), g(x2,⋅), . . . , g(xn,⋅)]∈∞×n, where xi is the ith sample in X, that is
In order to make every kernel-transformed sample in g(X,⋅) finite-dimensional, a further sample mapping manipulation is conducted on g(X,⋅) with the transformation matrix V, that is, g(X,⋅) is mapped to VTg(X,⋅). Here, the transformation matrix V is specifically designed as g(X,⋅)W, where W∈n×m
According to the characteristics of kernels, G=g(X,⋅)Tg(X,⋅)=[g(xi,⋅)Tg(xj,⋅)]1≤i,j≤n=[g(xi,xj)]1≤i,j≤n. In kernel embedding methods, researchers generally use predefined kernels, such as Gaussian kernel, Laplace kernel and polynomial kernel, etc. If the kernel g is determined g(xi,xj) is determined and easy to calculate based on a specific closed-form formula. Here we use Gaussian kernel, that is
where σ is the bandwidth parameter that needs to be tuned in real application. Therefore, in order to obtain a suitable transformed samples WTG, only W needs to be optimized. The goal of the sample transformation mechanism is to minimize the discrepancies of non-target samples from target samples while preserving the categorical knowledge of the samples. Hence, the problem of optimizing sample transformation can be formulated as
where tr(⋅) denotes the trace of matrix, Ω=Σc∈yΩc, Ωc=Σc∈yωcωcT, ωc=[ω1c, ω2c, . . . , ωnc]T,
ntary=c and ntary=c respectively are the number of target and non-target samples belonging to category c (c∈), yi is the ith element in Y. μ is a trade-off parameter.
Within the constraint (namely, the equation that follows “s. t.”), In is the nth order identity matrix, Hn is the nth order centering matrix defined as
1∈n×1 is the column vector with all 1's. =L+γIn, L is a kernel matrix of Y with respect to kernel l, that is, L=l(Y,⋅)Tl(Y,⋅)∈n×n, where l(Y,⋅)=[l(y1,⋅), l(y2,⋅), . . . , l(yn,⋅)]∈∞×n is the matrix to which Y is manned. γ is a trade-off parameter and γIn is employed to avoid the singularity of .
In this optimization formulation, tr(WTGΩGW) measures the sample distribution discrepancies between the transformed X(ntdr) and X(tardr) of the same category, tr(WTW) represents the matrix complexity of W, and tr(WTGHnHnGW) evaluates the correlation degree between the transformed samples WTG and their labels Y. Such correlation can be recognized as the retaining volume of the SS evaluation knowledge contained in non-target and target samples after being transformed. The optimized solution W*∈n×m
It is noted that if the non-target base evaluator trained by the transformed non-target samples WTg(X,⋅)Tg(X(ntdr),⋅) is used to evaluate an unclassified target sample x(tst), x(tst) needs to be transformed to the same sample space of WTg(X,⋅)Tg(X(ntdr),⋅) first, that is, x(tst) is transformed to WTg(X,⋅)Tg(x(tst),⋅). After that, the non-target base evaluator can be used to classify WTg(X,⋅)Tg(x(tst),⋅) to obtain the classification result of x(tst).
If the non-target samples belong to s non-target sectors, the abovementioned non-target sample learning procedures in
According to
Assume a target sample x(tst) needs to be evaluated. To achieve this, the confidence of each base evaluator baevj on evaluating x(tst) needs to be calculated (1≤j≤(s+1)T). First, transform x(tst) and the target sample set X(tar) into the sample space corresponding to baevj. Then, within this sample space, select ntarnear samples in X(tar) which are the nearest to the x(tst). After that, employ baevj to classify these ntarnear nearest samples and use the classification accuracy as the confidence of baevj (and also the weight of baevj) on evaluating x(tst). Thereupon, a confidence threshold θ is used to determine which base evaluators should be selected out to compose the final SS evaluator. At last, integrate the outputs of all the selected base evaluators through weighted voting to obtain the final result of evaluating x(tst).
The confidence threshold θ is an empirical parameter that is to be adjusted in real applications. The above steps of base evaluator integration will be implemented independently for every target sample that needs to be evaluated.
Number | Date | Country | Kind |
---|---|---|---|
201711071217.1 | Nov 2017 | CN | national |