Evaluation method and device of air traffic control speech recognition system

Information

  • Patent Grant
  • 12020688
  • Patent Number
    12,020,688
  • Date Filed
    Tuesday, December 26, 2023
    11 months ago
  • Date Issued
    Tuesday, June 25, 2024
    5 months ago
Abstract
An evaluation method and system of an air traffic control (ATC) speech recognition system are provided, the method includes: S1, obtaining, grading and inputting ATC speech data; S2, constructing an evaluation index system; S3, determining weights of utility layer indexes and support layer indexes under different levels ATC speech data; S4, calculating a utility layer score and a support layer score, and adding the scores of the utility layer and the support layer to obtain comprehensive scores of the utility layer and the support layer; S5, determining, weights of the utility layer and the support layer; S6, adding a product of the utility layer weight with the utility layer comprehensive score and a product of the support layer weight with the support layer comprehensive score to obtain a comprehensive score of the speech recognition system; and S7, determining a level of the recognition performance of the speech recognition system.
Description
TECHNICAL FIELD

The disclosure relates to the field of speech recognition, and more particularly to an evaluation method and system of an air traffic control (ATC) speech recognition system.


BACKGROUND

In recent years, with a continuous development of an aviation industry, a number of flights is also increasing year by year, which poses a more severe challenge to keep air traffic smooth and ensure flight safety. Communication and coordination between air traffic controllers and the controller sending a control instruction to crew are main means to adjust flight conflict, promote smooth air traffic, and ensure flight safety. Therefore, it is of great significance to correctly understand air traffic control (ATC) speech to ensure flight safety.


A speech recognition system can convert speech information to text information, thus increasing an information acquisition dimension to prevent an incorrect understanding for information between the controllers and between the controller and the crew, relieving workload of the controller, and ensuring the flight safety. An ATC instruction is a language with professional characteristics, main features of the ATC instruction include fast speech speed, concise and fixed instruction structure and special pronunciation. In the current stage, a quality of the speech recognition system is uneven, different speech recognition systems consider different emphasizes, such as response time, accuracy and other indexes, which makes performance and a recognition effect of the speech recognition system have a big difference, and there is a lack of a method for evaluating the quality of the speech recognition system to quantitatively evaluate the speech recognition system uniformly.


In a field of speech recognition, current evaluation methods are mainly aimed at voice quality and recognition algorithm, an evaluation method for the speech recognition system has not proposed yet. Based on a particularity of an ATC field, it is necessary to propose an evaluation method of an ATC speech recognition system in the ATC field based on traditional system evaluation methods, to overcome a problem of unified quantitative evaluation of the speech recognition systems in the ATC field.


SUMMARY

A purpose of the disclosure is to overcome a problem of unified quantitative evaluation of a speech recognition system in the related art, and provide an evaluation method and system of an air traffic control (ATC) speech recognition system.


In order to achieve the above purpose, the disclosure provides the following technical solutions.


An evaluation method of an ATC speech recognition system is provided, and the evaluation method includes:


S1, obtaining ATC speech data; grading, based on a quality level of the ATC speech data, the ATC speech data to obtain speech data with different quality levels, and inputting the speech data with different quality levels into a to-be-evaluated ATC speech recognition system;


S2, constructing an evaluation index system; where indexes of the evaluation index system include utility layer indexes and support layer indexes;


S3, determining, by using an improved group-G2 method and an improved criteria importance through intercriteria correlation (CRITIC) method, weights of the utility layer indexes and weights of the support layer indexes under the ATC speech data with different quality levels;


S4, calculating, according to the weights of the utility layer indexes and the weights of the support layer indexes under the ATC speech data with different quality levels, scores of a utility layer and scores of a support layer under the ATC speech data with different quality levels by using an improved technique for order preference by similarity to ideal solution (TOPSIS) method; and adding the scores of the utility layer to obtain a comprehensive score of the utility layer and adding the scores of the support layer to obtain a comprehensive score of the support layer;


where the TOPSIS method is to determine Z+ and Z of an index through a statistical method, to thereby score a recognition performance of the to-be-evaluated ATC speech recognition system; and


where Z+ represents a set containing optical values of evaluation indexes of the to-be-evaluated ATC speech recognition system, and Z represents a set containing worst values of the evaluation indexes of the to-be-evaluated ATC speech recognition system;


S5, determining, according to the comprehensive score of the utility layer and the comprehensive score of the support layer, a weight of the utility layer and a weight of the support layer of the to-be-evaluated ATC speech recognition system by using the improved group-G2 method and the improved CRITIC method;


S6, multiplying the weight of the utility layer and the comprehensive score of the utility layer to obtain a first product, multiplying the weight of the support layer and the comprehensive score of the support layer to obtain a second product, and adding the first product and the second product to obtain a comprehensive score of the to-be-evaluated ATC speech recognition system; and S7, determining, according to the comprehensive score of the to-be-evaluated ATC speech recognition system, a level of the recognition performance of the to-be-evaluated ATC speech recognition system.


In an exemplary embodiment, the evaluation method of the ATC speech recognition system further includes: comparing the level of the recognition performance of the to-be-evaluated ATC speech recognition system with levels of the recognition performance of other ATC speech recognition systems individually, thereby to obtain a target ATC speech recognition system, and applying the target ATC speech recognition system to recognize an ATC instruction, to thereby make an air traffic controller or a crew member correctly understand the ATC instruction.


In an embodiment, in step S2, the utility layer indexes include accuracy indexes and a recognition rate index, the recognition rate index includes a processing time consumption, and the processing time consumption refers to a time consumed from a moment when the to-be-evaluated ATC speech recognition system receives completely inputted ATC speech data to a moment when the to-be-evaluated ATC speech recognition system recognizes a result and outputs the result.


In an embodiment, the accuracy indexes include a sentence accuracy, a word accuracy and a keyword accuracy, and a calculation expression of the keyword accuracy is expressed as follows:








K

a

c

c


=

1
-

(



K

e

r

r

o

r



K

t

o

t

a

l



×
100

%

)



;




where Kacc represents the keyword accuracy, Kerror represents a number of keyword errors in a sentence, and Ktotal represents a total number of keywords in a sentence.


In an embodiment, in step S2, the support layer indexes include reliability indexes and maintainability indexes.


In an embodiment, in step S4, calculation formulas of Z+ and Z of the index in the improved TOPSIS method are as follows:

Z+=(max{z11,z21, . . . ,zn1},max{z12,z22, . . . ,zn2}, . . . max{z1n,z2n, . . . ,znn};
Z=(min{z11,z21, . . . ,zn1},min{z12,z22, . . . ,zn2}, . . . min{z1n,z2n, . . . ,znn};


where z11 represents a first index value of a first system, z21 represents a first index value of a second system, z22 represents a second index value of a second system, z2n represents a n-th index value of the second system, and znn represents a n-th index value of a n-th system.


In an embodiment, in step S7, the determining, according to the comprehensive score of the to-be-evaluated ATC speech recognition system, a level of the recognition performance of the to-be-evaluated ATC speech recognition system includes: determining, according to the comprehensive score of the to-be-evaluated ATC speech recognition system and a determination principle, the level of the recognition performance of the to-be-evaluated ATC speech recognition system; and the determination principle includes:


determining, in response to the comprehensive score being in a range of 0.0-0.3 scores, containing 0.3 scores, the level of the recognition performance to be level 1;


determining, in response to the comprehensive score being in a range of 0.3-0.5 scores, containing 0.5 scores, the level of the recognition performance to be level 2;


determining, in response to the comprehensive score being in a range of 0.5-0.7 scores, containing 0.7 scores, the level of the recognition performance to be level 3;


determining, in response to the comprehensive score being in a range of 0.7-0.9 scores, containing 0.9 scores, the level of the recognition performance to be level 4; and


determining, in response to the comprehensive score being in a range of 0.9-1.0 scores, containing 1.0 score, the level of the recognition performance to be level 5.


An evaluation system of the ATC speech recognition system is provided, the evaluation system includes at least one processor and a memory communicatively connected to the at least one processor, the memory stores an instruction executed by the at least one processor, and the instruction is configured to be executed by the at least one processor, to make the at least one processor to execute any one steps of the evaluation method.


Compared to the related art, beneficial effects of the disclosure are as follows.

    • 1. The improved TOPSIS method overcomes the problem of unified quantitative evaluation of the speech recognition system in an ATC field, and achieves comprehensive quantification evaluations of single and multiple indexes for the ATC speech recognition system.
    • 2. A method fusing the improved group-G2 method and the improved CRITIC method of the disclosure is used to determine a weight of each index, the method is a specific implementation of subjective weighting combined with objective weighting to ultimately determine the weights of the indexes, and the method overcomes one-sidedness caused by subjective weighting and a deviation caused by objective weighting, and makes an evaluation effect more reliable.





BRIEF DESCRIPTION OF DRAWINGS


FIG. 1 illustrates a flowchart of an evaluation method of an air traffic control (ATC) speech recognition system according to an embodiment of the disclosure.



FIG. 2 illustrates a tree diagram of indexes of the evaluation method of the ATC speech recognition system according to an embodiment of the disclosure.



FIG. 3 illustrates a schematic diagram of working times of the ATC speech recognition system according to an embodiment of the disclosure.



FIG. 4 illustrates a flowchart of determination of weights of utility layer indexes and weights of support layer indexes according to an embodiment of the disclosure.





DETAILED DESCRIPTION OF EMBODIMENTS

The disclosure will be described in detail below in conjunction with embodiments and specific implementations. However, it cannot be understood that a scope of the above topics of the disclosure is merely limited to the following embodiments, and all technologies implemented based on a content of the disclosure fall within a scope of the disclosure.


Embodiment

As shown in FIG. 1, the embodiment includes the following steps S1-S7.


In step S1, air traffic control (ATC) speech data is obtained, the ATC speech data is graded to obtain ATC speech data with different quality levels, and the ATC speech data with different quality levels is inputted into a to-be-evaluated ATC speech recognition system.


In step S2, an evaluation index system is constructed, and indexes of the evaluation index system include utility layer indexes and support layer indexes.


In step S3, weights of the utility layer indexes and weights of the support layer indexes under the ATC speech data with different quality levels are determined by using an improved group-G2 method and an improved criteria importance through intercriteria correlation (CRITIC) method.


In step S4, scores of a utility layer and scores of a support layer under the ATC speech data with different quality levels are calculated according to the weights of the utility layer indexes and the weights of the support layer indexes under the ATC speech data with different quality levels and by using an improved technique for order preference by similarity to ideal solution (TOPSIS) method, the scores of the utility layer are added to obtain a comprehensive score of the utility layer, and the scores of the support layer are added to obtain a comprehensive score of the support layer. The TOPSIS method is to determine Z+ and Z of an index through a statistical method, to thereby score a recognition performance of the to-be-evaluated ATC speech recognition system. Z+ represents a set containing target values of evaluation indexes of the to-be-evaluated ATC speech recognition system, and Z represents a set containing worst values of the evaluation indexes of the to-be-evaluated ATC speech recognition system.


In step S5, a weight of the utility layer and a weight of the support layer of the to-be-evaluated ATC speech recognition system are determined according to the comprehensive score of the utility layer and the comprehensive score of the support layer and by using the improved group-G2 method and the improved CRITIC method.


In step S6, the weight of the utility layer and the comprehensive score of the utility layer are multiplied to obtain a first product, the weight of the support layer and the comprehensive score of the support layer are multiplied to obtain a second product, and the first product and the second product are added to obtain a comprehensive score of the to-be-evaluated ATC speech recognition system.


In step S7, a level of the recognition performance of the to-be-evaluated ATC speech recognition system is determined according to the comprehensive score of the to-be-evaluated ATC speech recognition system.


The improved group-G2 method is a subjective weighting method with high flexibility and can largely reflect risk awareness characteristics of experts, and the improved group-G2 method includes the following steps (1)-(4).


In step (1), a weight of a minimum class non empty intersection (i.e., a comprehensive assignment interval of an expert for indexes) is calculated as follows.

    • 1) The minimum class non empty intersection.


xi represents an evaluation index, xm represents a least important index recognized by the experts, ai(k) represents a ratio of importance degree between the index xi and the index xm by an expert k. Di(k) represents an assignment interval of ai(k), where d1i(k) and d2i(k) are respectively left and right end points of the interval, for the index m, d1m(k)=d2m(k)=1 (the least important index is set as 1). Therefore, a minimum class non empty intersection of Di(k) is set as DiA, let a set S={1, 2, 3, . . . , L (L experts)}, and the minimum class non empty intersection DiA should satisfy the following conditions 1-3.


In condition 1, A represents a non-empty subset of S.


In condition 2, DiA=∩k∈A Di(k)≠Ø.


In condition 3, w∈S exists, and w∉A, and all of this satisfy DiA ∩Di(w)=Ø.

    • 2) Adjacency degree: n is set as a number of the minimum class non empty intersections of an i-th index, DiAj is recorded as a j-th minimum class non empty intersection for the i-th index, xxij and xxij+ are respectively left and right end points of DiAj, Ij represents an adjacency degree between the j-th minimum class non empty intersection and other minimum class non empty intersections, and calculation formulas of the adjacency degree are expressed as follows:











I
j

=

1
/







p
=
1

,

p

j


n






D
i

A
j


-

D
i

A
p







;










D
i

A

j


-

D
i

A

p





=



[



(


x

x

i

j


-

-

x

i

p

-


)

2

+


(


x

x

i

j


+

-

x

i

p

+


)

2


]


1
/
2


.










    • 3) Capacity: Pj is set as a capacity of DiAj, that is, Aj represents a number of intervals obtained by how many intervals intersect.

    • 4) A calculation formula of a weight of the minimum class non empty intersection is expressed as follows:











v
j

=


P
j



1
j








i
=
1

n



P
i



1
i



,

j
=
1

,
2
,
3
,


,

n
;





where vj represents a weight of the j-th minimum class non empty intersection.


In step (2), the comprehensive assignment interval of the expert for the index is calculated as follows.


A formula of the comprehensive assignment interval of the group experts on a comprehensive importance of an index is expressed as follows:








D
i

=







j
=
1

n



v
j



D
i

A

j




,

i
=
1

,
2
,


,

m
;





where n represents a number of the minimum class non-empty intersections, and m represents a number of indexes.


In step (3), a risk attitude factor is determined as follows.


Ik* represents an adjacency degree between the assignment interval Di(k) of the ratio of importance degree between the index xi and the index xm by the expert and other assignment intervals, and calculation formulas of Ik* are expressed as follows:











I
k
*

=

1
/







j
=
1

,

j

k


n






D
i

(
k
)


-

D
i

(
j
)







,

k
=
1

,
2
,


,

L
;









r
k

=


I
k
*

/






j
=
1

n



I
j
*



,

k
=
1

,
2
,


,

L
;








where rk represents a result of normalizing the adjacency degree of the assignment interval of the k-th expert.


A calculation formula of a risk attitude factor λi is expressed as follows:








λ
i

=


[



-






h
=
1


Q
i
L






L


S
i
h



d
1




r

N

h



+







h
=
1


Q
i
R





R


S
i
h



d
2




r

M

h




]


1
2



;




where rNh represents a normalized adjacency degree of an assignment interval of a Nh-th expert, rMh represents a normalized adjacency degree of an assignment interval of a Mh-th expert, rNh=(1, 2, . . . , QiL) and rMh=(1, 2, . . . QiR), which correspond to rk(k=1, 2, 3, . . . , L) one-by-one.


Relevant concepts involved in the above formula are as follows.

    • 1) n(Di(k)) represents a middle point of the assignment interval Di(k); and QiL represents a number of middle points on a left of n(Di).
    • 2) n(Di(Nh)) represents a h-th middle point on an assignment interval to the left of n(Di) from left to right.
    • 3) QiR represents a number of middle points on a right of n(Di), and contains n(Di); and QiL+QiR=L.
    • 4) n(Di(Mh)) represents a h-th middle point on an assignment interval to the right of n(Di) from left to right, and contains a middle point of the interval n(Di).
    • 5) LSih represents a distance between n(Di(Nh)) and n(Di), RSih represents a distance between n(Di(Mh)) and n(Di), LSih/d1 represents a left deviation distance between n(Di(Nh)) and n(Di), and h=1, 2, 3, . . . , QiL; RSih/d2 represents a right deviation distance between n(Di(Mh)) and n(Di); and d1=n(Di)−n(Di(N1)).
    • 6) n(Di(N1)) represents a middle point on the left of n(Di) and facing away from n(Di); d2=n(Di(Mh))−n(Di), and h=QiR; and n(Di(Mh)) represents a middle point on the right of n(Di) and facing away from n(Di).


In step (4), a weight of the index is determined as follows.


ƒλi is set as an interval mapping function of the group experts for the risk attitude factor λi; and e(Di) represents a length of the interval Di, and a formula of ƒλi is expressed as follows:

ƒλi(Di)=n(Di)+λie(Di).


wi is set as a weight of the i-th index, and a formula of wi is expressed as follows:







w
i

=



f

λ
i


(

D
i

)

/






j
=
1

m





f

λ
i


(

D
i

)

.






The improved CRITIC method is an objective weighting method, an existing entropy weighting method only considers an information entropy value of data, while an existing CRITIC method only considers fluctuation and conflict of data. Therefore, the improved CRITIC method is needed, the improved CRITIC method fully considers the fluctuation, the conflict and the information entropy value of the data, and the improved CRITIC method includes the following steps (1)-(7).


In step (1), an evaluation index matrix is constructed.


In step (2), data is standardized, which mainly makes performance normalized and standardized.


In step (3), a proportion of an i-th object of a j-th index is calculated, and a calculation formula is expressed as follows:







P

i

j


=



x

i

j









i
=
1

m



x

i

j




.





In step (4), an information entropy of the j-th index is calculated, and a calculation formula is expressed as follows:







e
j

=


-

1

ln


n










i
=
1

m



P

i

j



ln



P

i

j


.






In step (5), a standard deviation of data is calculated, and calculation formulas are expressed as follows:












x
¯

j

=


1
m








i
=
1

m



x

i

j




;







S
j

=










i
=
1

m




(


x

i

j


-


x
¯

j


)

2



n
-
1



.








In step (6), a correlation matrix of the data is calculated, and a calculation formula is expressed as follows:






R
=









j
,

k
=
1


n



(


x

i

j


-


x
¯

j


)



(


x

i

k


-


x
¯

k


)









j
=
1

n




(


x

i

j


-


x
¯

j


)

2








k
=
1

n




(


x

i

k


-


x
¯

k


)

2



.





In step (7), a weight is calculated, and a calculation formula is expressed as follows:








W
j

=



(


e
j

+

S
j


)








i
=
1

n



(

1
-



"\[LeftBracketingBar]"


r

i

j




"\[RightBracketingBar]"



)









j
=
1

n



(


e
j

+

S
j


)








i
=
1

n



(

1
-



"\[LeftBracketingBar]"


r

i

j




"\[RightBracketingBar]"



)




;




where rij represents a correlation coefficient of the i-th index and the j-th index.


A fusion formula of a subjective weight and an objective weight is expressed as follows:








W
k

=



w

o

k




w

s

k










k
=
1

m



w

o

k




w

s

k





;




where wok represents an objective weight of an index k, and wsk represents a subjective weight of the index k.


An introduction of the improved TOPSIS method is as follows.


(1) The TOPSIS method before improving includes the following steps 1) to 2).


In step 1), normalizations of the indexes are as follows.


In the utility layer, accuracy indexes are extremely large indexes, a processing time consumption index is an extremely small index, and it is necessary to convert the processing time consumption index to an extremely large index. In the support layer, a maintainability exponent, a mean time between failures (MTBF) and a mean time to failure (MTTF) are extremely large indexes, and a module coupling degree and a mean time to repair (MTTR) are extremely small indexes, and it is necessary to convert the module coupling degree and the mean time to repair to extremely large indexes.


In step 2), standardization of the data is as follows.


In order to eliminate effects of dimensions of different indexes, it is necessary to normalize and standardize the index data matrix, thus obtaining a usable evaluation index matrix, a score of each evaluation object is calculated according to a scoring formula, and the score is calculated based on the following distance formula:







score


of


x

=



distance


between


x


and


minimum



distance


between


x


and


maximum

+

distance


between






x


and


minimum



.





Specifically, a maximum and a minimum of an index are found and recorded as Zj+ and Zj, and a distance formula of a j-th index of an i-th evaluation object and the maximum is expressed as follows:








D
i
+

=








j
=
1

m





w
j

(


Z
j
+

-

z

i

j



)

2




;




where Zj+ represents a maximum of the j-th index, and zij represents a j-th normalized and standardized index value of the i-th evaluation object.


Similarly, a distance formula of the j-th index of the i-th evaluation object and the minimum is expressed as follows:








D
i
-

=








j
=
1

m





w
j

(


Z
j
-

-

z

i

j



)

2




;




where wj represents a weight of the index j; and a score Si of the i-th evaluation object is calculated as follows:







S
i

=



D
i
-



D
i
+

+

D
i
-



.





(2) The improved TOPSIS method includes the following steps 1) to 3).


In step 1), indexes are normalized.


In step 2), data is standardized.


In step 3), a score is calculated.


Formulas proposed in the TOPSIS method before improving is only applicable when there are multiple objects to be evaluated, calculation results of different batches of evaluation objects may not be comparable due to a lack of unified standards during calculation, and scoring results vary greatly when calculated in different batches for a same system.


In the embodiment, determination formulas of Z+ and Z of the index determined by using the improved TOPSIS method and a statistical method are expressed as follows:

Z+=(max{z11,z21, . . . ,zn1},max{z12,z22, . . . ,zn2}, . . . max{z1n,z2n, . . . ,znn};
Z=(min{z11,z21, . . . ,zn1},min{z12,z22, . . . ,zn2}, . . . min{z1n,z2n, . . . ,znn};


For evaluating the speech recognition system, the index value is fixed, Z+ and Z of the index are determined by the statistical method, and a recognition performance of each to-be-evaluated objects is scored by using the statistical method, so as to achieve scoring for a single object, and the to-be-evaluated objects of different batches have comparability due to unified standards.


A level of the recognition performance of the speech recognition system is graded according to the comprehensive score of the speech recognition system obtained by the method of the disclosure, which are divided into levels 1-5, and level 5 is the highest level. In response to the comprehensive score being a range of 0.0-0.3 scores, containing 0.3 scores, a level of the recognition performance is level 1; in response to the comprehensive score being a range of 0.3-0.5 scores, containing 0.5 scores, a level of the recognition performance is level 2; in response to the comprehensive score being a range of 0.5-0.7 scores, containing 0.7 scores, a level of the recognition performance is level 3; in response to the comprehensive score being a range of 0.7-0.9 scores, containing 0.9 scores, a level of the recognition performance is level 4; and in response to the comprehensive score being a range of 0.9-1.0 scores, containing 1.0 score, a level of the recognition performance is level 5.


A specific recognition performance of each level is as follows.


When the level of the recognition performance is level 1, the recognition system has a low vocabulary and a slow recognition speed, a recognition result does not meet a standard specification, the recognition system can only effectively recognize specific input audio, robustness of the recognition system is poor, and the recognition performance cannot meet requirements.


When the level of the recognition performance is level 2, the recognition system has a limited vocabulary and a slow recognition speed, a grammar structure of a recognition result deviates significantly from the standard specification, the recognition system is more susceptible to interference from audio, resulting in significant deviations in the recognition result, and requires a high standard for input speech, resulting in a poor overall recognition performance.


When the level of the recognition performance is level 3, a vocabulary of the recognition system meets a basic requirement, a grammar structure of a recognition result basically meets the standard specification, a recognition speed is normal, some insertion and deletion errors occur occasionally, and some words are ambiguous.


When the level of the recognition performance is level 4, a vocabulary of the recognition system has a wide coverage, a grammar structure of a recognition result meets the standard specification, a recognition speed is fast, and some words are ambiguous.


When the level of the recognition performance is level 5, a vocabulary of the recognition system has a wide coverage, a grammar structure of a recognition result meets the standard specification, a recognition speed is fast, and recognition is correct and is not ambiguous.


In an embodiment, in step 1, the obtained ATC speech data is graded to 5 levels according to a mean opinion score (MOS) terminology, and a gradation state is shown in Table 1.









TABLE 1







An evaluation standard for grading the ATC speech data









Audio level
MOS value
Evaluation standard





Excellent
4.0-5.0
Excellent, listen clearly, low latency,




and smooth communication


Good
3.5-4.0
Second best, listen clearly,




low latency, lack of smooth




communication, and a bit of noise


Medium
3.0-3.5
Not bad, unable to listen clearly,




a certain delay, and be able to




communicate


Poor
1.5-3.0
Reluctant, cannot listen clearly,




large delay, and communication




requires multiple repetitions


Bad
  0-1.5
Extremely poor, unintelligible,




large delay, and bad communication









In an embodiment, in step 2, the utility layer indexes include the accuracy indexes and a recognition rate index, the accuracy indexes include a sentence accuracy, a word accuracy and a keyword accuracy, and the utility layer indexes are shown in FIG. 2.


An expression for calculating the sentence accuracy is expressed as follows:








S

a

c

c


=

1
-

(



E
s


S
t


×
1

0

0

%

)



;




where Sacc represents the sentence accuracy, Es represents a number of misrecognized sentences, and St represents a total number of recognized sentences.


An expression for calculating the word accuracy is expressed as follows:








W

a

c

c


=


T
N

×
1

0

0

%


;




where Wacc represents the word accuracy, T represents a number of correctly recognized words, and N represents a total number of words.


An expression for calculating the keyword accuracy is expressed as follows:








K

a

c

c


=

1
-

(



K

e

r

r

o

r



K

t

o

t

a

l



×
100

%

)



;




where Kacc represents the keyword accuracy, Kerror represents a number of keyword errors in a sentence, and Ktotal represents a total number of keywords in a sentence


The recognition rate index includes a processing time consumption, and the processing time consumption refers to a time consumed from a moment when the ATC speech recognition system receives completely inputted ATC speech data to a moment when the ATC speech recognition system recognizes a result and outputs the result.


In an embodiment, in step S2, the support layer indexes include reliability indexes and maintainability indexes.


The reliability indexes include a MTTF, a MTTR and a MTBF.


As shown in FIG. 3, the MTTF refers to a mean time for the speech recognition system to operate without failures, which is a mean time of time periods between a start of normal operation and an occurrence of a malfunction of the speech recognition system, and a calculation formula is expressed as follows:







MTTF
=




T
1


n


;




where T1 represents a time to operate without failures, and n represents a number of time periods between the start of normal operation and the occurrence of the malfunction of the speech recognition system.


The MTTR refers to a mean time of time periods between the occurrence of the malfunction and an end of maintenance of the speech recognition system, and a calculation formula is expressed as follows:







MTTR
=




(


T
2

+

T
3


)


n


;




where T2 represents a time taken from the start of the malfunction to a start of maintenance, T3 represents a time of maintenance, and n represents a number of time periods between the occurrence of the malfunction and the end of maintenance.


The MTBF refers to a mean time of time periods between two failures of the speech recognition system, and a calculation formula is expressed as follows:







MTBF
=




(


T
2

+

T
3

+

T
1


)


n


;




where T1 represents the time to operate without failures, T2 represents the time taken from the start of the malfunction to a start of maintenance, T3 represents the time of maintenance, and n represents a number of time periods between two failures.


The maintainability indexes include a maintainability exponent and a module coupling degree.


A calculation formula of the maintainability exponent is expressed as follows:

171−5.2 ln(HV)−0.23CC−16.2 ln(LOC)+50 sin √{square root over (2.46×COM)};


where HV represents a Halstead value, CC represents a cyclomatic complexity, LOC represents an average number of code lines for each module, and COM represents a percentage of comment lines for each module.


A calculation formula of the module coupling degree is expressed as follows:







Coupling



(
C
)

=

1
-

1


d
i

+

2


c
i


+

d
o

+

2


c
o


+

g
d

+

2


g
c


+
r
+
w





;




where di represents a number of inputted data parameters, ci represents a number of inputted control parameters, do represents a number of outputted data parameters, co represents a number of global couplings of outputted control parameters, gd represents a global variable for storing data, gc represents a global variable for controlling environmental coupling, w represents a number of fan-out modules called by the module, and r represents a number of fan-in modules calling the modules.


In an embodiment, in step S3, the weights of the utility layer indexes and the weights of the support layer indexes under the ATC speech data with different quality levels are determined by using the improved group-G2 method and the improved CRITIC method, and step 3 specifically includes the following steps S31-S33.


In step S31, a subjective weight of each utility layer index is calculated by using the improved group-G2 method, the improved group-G2 method is a subjective weighting method for inflecting risk awareness characteristics of the experts, and formulas are expressed as follows:











w
i

=


f

λ
i




(

D
i

)

/






j
=
1

m



f

λ
i




(

D
j

)



;









f

λ
i




(

D
i

)


=


n


(

D
i

)


+


λ
i


e


(

D
i

)




;







where ƒλi represents the interval mapping function of the group experts for the risk attitude factor λi, e(Di) represents a length of the interval Di, and wi represents the subjective weight of the i-th index.


In step S32, an objective weight of each utility layer index under the ATC speech data with different quality levels is calculated by using the improved CRITIC method, the improved CRITIC method is an objective weighting method for the fluctuation, the conflict and the information entropy value of the data, and formulas are expressed as follows:











W
j

=



(


e
j

+

S
j


)








i
=
1

n



(

1
-



"\[LeftBracketingBar]"


r

i

j




"\[RightBracketingBar]"



)









j
=
1

n



(


e
j

+

S
j


)








i
=
1

n



(

1
-



"\[LeftBracketingBar]"


r

i

j




"\[RightBracketingBar]"



)




;








e
j

=


-

1

ln

n










i
=
1

m



P

i

j



ln


P

i

j




;








S
j

=









i
=
1

m




(


x

i

j


-


x
¯

j


)

2



n
-
1




;









x
¯

j

=


1
m








i
=
1

m



x

i

j




;








P

i

j


=


x

i

j









i
=
1

m



x

i

j





;







where Pij represents the proportion of the i-th object of the j-th index, ej represents the information entropy value of the j-th index, Sj represents a standard deviation of the data, rij represents a correlation coefficient of the i-th index and the j-th index, and Wj represents an objective weight.


In step S33, the weight obtained by the subjective weighting method (i.e., the improved group-G2 method) and the weight obtained by the objective weighting method (i.e., the improved CRITIC method) are fused to obtain a comprehensive weights of each utility layer index and a comprehensive weight of each support layer index under the ATC speech data with different quality levels, and a formula is expressed as follows:








W
k

=



w

o

k




w

s

k










k
=
1

m



w

o

k




w

s

k





;




where wok represents the objective weight of the index k, and wsk represents the subjective weight of the index k.


The evaluation method of the ATC speech recognition system is adopted in the embodiment, four speech recognition systems are actually simulated for instance validation by using a Matlab 2020 software, a system 1, a system 2 and a system 3 are used to calculate weights, and a system 4 is used to final score. Performance of the utility layer indexes and the support layer indexes of each system under the ATC speech data with five quality levels are listed as follows.


The performance of the utility layer indexes of the systems under the ATC speech data with different quality levels is as follows.









TABLE 2







Performance of the utility layer indexes of the systems


under level 1 ATC speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)














System 1
85%
  87%
  89%
1.8


System 2
88%
88.5%
89.5%
2


System 3
89%
  90%
  90%
2.1


System 4
91%
91.5%
  92%
1.8
















TABLE 3







Performance of the utility layer indexes of the systems


under level 2 ATC speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)














System 1
87%
  88%
90%
1.6


System 2
89%
89.5%
91%
1.8


System 3
90%
  91%
91%
1.9


System 4
92%
  93%
93%
1.6
















TABLE 4







Performance of the utility layer indexes of the systems


under level 3 ATC speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)














System 1
  89%
89.5%
  90%
1.5


System 2
90.5%
  91%
  91%
1.74


System 3
  91%
  90%
91.5%
1.87


System 4
  94%
  94%
  95%
1.3
















TABLE 5







Performance of the utility layer indexes of the systems


under level 4 ATC speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)














System 1
92%
  92%
93%
1.35


System 2
93%
  93%
94%
1.45


System 3
94%
94.5%
95%
1.55


System 4
95%
95.5%
96%
1.2
















TABLE 6







Performance of the utility layer indexes of the systems


under level 5 ATC speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)














System 1
93%
  94%
94%
1.3


System 2
94%
94.5%
95%
1.5


System 3
95%
  96%
96%
1.6


System 4
97%
  97%
97%
1.2









Performance of the support layer indexes of the systems is as follows.









TABLE 7







Performance of the support layer indexes of the systems













Maintainability
Module coupling
MTTF
MTTR
MTBF



exponent
degree
(h)
(h)
(h)















System 1
89
0.78
8000
30
8030


System 2
80
0.82
8300
36
8330


System 3
75
0.95
7200
401
7240


System 4
90
0.3
13000
20
13020









The improved group-G2 method and the improved CRITIC method are combined to calculate the subjective weight and the objective weight of each utility layer index under the ATC speech data with different quality levels, and the subjective weight and the objective weight are fused to obtain the comprehensive weight of each utility layer index under the ATC speech data with different quality levels; and a same method is used to determine the comprehensive weight of each support layer index.


The improved group-G2 method is used to calculate the weight of each utility layer index.









TABLE 8







Weights of the utility layer indexes under 5 levels ATC speech data calculated by


using the improved group-G2 method



















Minimum







Assignment
Assignment
Assignment
class non

Risk
Interval




interval
interval
interval
empty
Comprehensive
attitude
mapping
Index


Index
Di(1) of an
Di(2) of an
Di(3) of an
intersection
assignment
factor
function
weight


xi
expert 1
expert 2
expert 3
DiAJ
interval
λi
ƒλi
wi


















Index
[1, 1]
[1, 1]
[1, 1]
[1, 1]
[1, 1]
0
1
0.106


1










Index
[1.5, 2.5]
[2, 3]
[1.5, 3]
[2, 2.5]
[2, 2.5]
0
2.25
0.238


2










Index
[1.5, 2.5]
[1.5, 2.5]
[1.5, 2.5]
[2.5, 3]
[2.5, 3]
−0.6
2.45
0.259


3










Index
[1.5, 2.5]
[1.5, 2.5]
[1.5, 2.5]
[3.5, 4]
[3.5, 4]
0
3.75
0.397


4

















Specifically, the index 1 is the sentence accuracy; the index 2 is the word accuracy; the index 3 is the keyword accuracy; and the index 4 is the processing time consumption.


Similarly, the weights of the utility layer indexes under 4 levels ATC speech data obtained by using the improved group-G2 method are shown in Table 9.









TABLE 9







The weights of the utility layer indexes under the 4 levels ATC


speech data obtained by using the improved group-G2 method












Weight
Weight
Weight
Processing


Level of
of the
of the
of the
time


the ATC
sentence
word
keyword
consumption


speech data
accuracy
accuracy
accuracy
(s)














Level 1 ATC
0.17
0.295
0.34
0.195


speech data






Level 2 ATC
0.16
0.29
0.32
0.23


speech data






Level 3 ATC
0.14
0.26
0.3
0.3


speech data






Level 4 ATC
0.104
0.242
0.286
0.368


speech data
















TABLE 10







Weights of the support layer indexes calculated by using the improved group-G2 method



















Minimum







Assignment
Assignment
Assignment
class non

Risk
Interval




interval
interval
interval
empty
Comprehensive
attitude
mapping
Index


Index
Di(1) of an
Di(2) of an
Di(3) of an
intersection
assignment
factor
function
weight


xi
expert 1
expert 2
expert 3
DiAJ
interval
λi
ƒλi
wi


















Index
[1, 1]
[1, 1]
[1, 1]
[1, 1]
[1, 1]
0
1
0.062


1










Index
[2, 3]
[2.5, 3]
[2.6, 3.6]
[2.6, 3]
[2.6, 3]
−0.0875
2.66
0.164


2










Index
[4, 6]
[3, 6]
[5, 6.5]
[5, 6]
[5, 6]
−0.222
5.278
0.326


3










Index
[3, 4]
[3, 4.5]
[2, 3.5]
[3, 3.5]
[3, 3.5]
0.184
3.342
0.207


4










Index
[3,5]
[4, 5.5]
[3,5]
[4, 5]
[4, 5]
−0.6
3.9
0.2411


5

















Weights of the utility layer indexes and the support layer indexes calculated by using the improved CRITIC method are as follows.


The weights of the utility layer indexes under the ATC speech data with different quality levels calculated by using the improved CRITIC method are shown in Table 11-15.









TABLE 11







Weights of the utility layer indexes under the level 1 ATC speech data calculated


by using the improved CRITIC method













Sentence

Word
 Keyword
Processing time



accuracy

accuracy
 accuracy
consumption (s)















Information
0.6290196609534563

05881199233394943
 0.5881199233394943
0.5220605575349556


entropy







Information
0.88142193

0.82515617
 0.82515617
2.47773866


carrying







capacity







Objective
0.26374510

0.25014913
 0.25014913
0.23595663


weight









The information carrying capacity is obtained by multiplying conflict and volatility of data.









TABLE 12







Weights of the utility layer indexes under the level 2 ATC speech data calculated


by using the improved CRITIC method















Processing time



Sentence accuracy
Word accuracy
Keyword accuracy
consumption (s)





Information
0.6203819994751361
0.5881199233394943
0.6374906669220128
0.5220605575349553


entropy






Information
0.86187345
0.87119148
1.00596527
2.46404423


carrying






capacity






Objective
0.25637963
0.24340510
0.26816446
0.23205081


weight
















TABLE 13







Weights of the utility layer indexes under the level 3 ATC speech data calculated


by using the improved CRITIC method















Processing time



Sentence accuracy
Word accuracy
Keyword accuracy
consumption (s)














Information
0.6290196609534563
0.5220605575349555
0.6203819994751361
0.5316247532311797


entropy






Information
1.02890285
1.00022968
1.04098174
2.26830129


carrying






capacity






Objective
0.27816315
0.20498861
0.27058092
0.24626732


weight
















TABLE 14







Weights of the utility layer indexes under the level 4 ATC speech data calculated


by using the improved CRITIC method

















Processing time



Sentence accuracy
Word accuracy
Keyword accuracy

consumption (s)















Information
0.5881199233394949
0.554099538248139
0.5881199233394933

05881199233394943


entropy







Information
0.81919132
0.82463451
0.81919132

2.44679501


carrying







capacity







Objective
0.25208077
0.24375768
0.25208077

0.25208077


weight
















TABLE 15







Weights of the utility layer indexes under the level 5 ATC speech data calculated


by using the improved CRITIC method















Processing time



Sentence accuracy
Word accuracy
Keyword accuracy
consumption (s)





Information
0.5881199233394933
0.46676943658628967
0.5881199233394943
0.5220605575349556


entropy






Information
0.82515617
0.83687839
0.82515617
2.43415730


carrying






capacity






Objective
0.26240273
0.23095275
0.26240273
0.24424179


weight









Weights of the support layer indexes calculated by using the improved CRITIC method is shown in Table 16.









TABLE 16







Weights of the support layer indexes calculated by using the improved CRITIC


method













Maintainability
Module coupling






exponent
degree
MTTF (h)
MTTR (h)
MTBF (h)





Information
0.5345474455983533
0.6301647186683301
0.6270410695481976
0.5540995382481381
0.6264291625127868


entropy







Information
0.34189017
0.15893173
0.33769637
0.30453009
0.34240899


carrying







capacity







Objective
0.18331129
0.21888241
0.20460045
0.18926360
0.20394226


weight









Comprehensive weights of the utility layer indexes are calculated as follows.









TABLE 17







Dynamic comprehensive weights of the utility layer indexes under levels 1-5 speech


data












Weight of the






sentence
Weight of the word
Weight of the
Processing time



accuracy
accuracy
keyword accuracy
consumption (s)





Level 1
0.17956724308621672
0.2955390047654
0.340621225831423
0.18427252631686084


speech data






Level 2
0.16356444213165897
0.28145765773482057
0.3421657976449812
0.21281210248853927


speech data






Level 3
0.15747565905989744
0.2155206467724271
0.3282496161954392
0.2987540779722362


speech data






Level 4
0.10483767901836524
0.23589460869306872
0.2883036173005044
0.3709640949880616


speech data






Level 5
0.11228833302739928
0.22190164162571346
0.2743648891895888
0.3914451361572985


speech data
















TABLE 18







Fixed comprehensive weights of the support layer indexes











Maintainability
Module coupling





exponent
degree
MTTF (h)
MTTR (h)
MTBF (h)





0.15132828144297342
0.06831083432737761
0.33574629044719095
0.1972079780724709
0.24740661570998718









An optimal value and a worst value of each index obtained from simulation statistics are substituted into the improved TOPSIS method to calculate scores of the utility layers and the support layers of the system 1, the system 2 and the system 3 under 5 level speech data, and a fixed comprehensive weight of the utility layer and a fixed comprehensive weight of the support layer are calculated by using the improved group-G2 method and the improved CRITIC method.


Optimal values and worst values of the utility layer indexes under level 1 speech data are shown in Table 19.









TABLE 19







Optimal values and worst values of the utility layer indexes


under level 1 speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)














Optimal value
96%
96%
96%
0.9


Worst value
80%
80%
80%
5









Optimal values and worst values of the utility layer indexes under level 2 speech data are shown in Table 20.









TABLE 20







Optimal values and worst values of the utility layer indexes


under level 2 speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)





Optimal value
97%
97%
97%
0.8


Worst value
82%
82%
82%
4.5









Optimal values and worst values of the utility layer indexes under level 3 speech data are shown in Table 21.









TABLE 21







Optimal values and worst values of the utility layer indexes


under level 3 speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)














Optimal value
98%
98%
98%
0.7


Worst value
83%
83%
83%
4









Optimal values and worst values of the utility layer indexes under level 4 speech data are shown in Table 22.









TABLE 22







Optimal values and worst values of the utility layer indexes


under level 4 speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)





Optimal value
99%
99%
99%
0.6


Worst value
85%
85%
85%
3.5









Optimal values and worst values of the utility layer indexes under level 5 speech data are shown in Table 23.









TABLE 23







Optimal values and worst values of the utility layer indexes


under level 5 speech data












Sentence
Word
Keyword
Processing time



accuracy
accuracy
accuracy
consumption (s)














Optimal value
99.5%
99.5%
99.5%
0.5


Worst value
  87%
  87%
  87%
3
















TABLE 24







Optimal values and worst values of the support layer indexes














Module






Maintainability
coupling
MTTF
MTTR
MTBF



exponent
degree
(h)
(h)
(h)















Optimal value
99
0.1
15000
10
15010


Worst value
60
1
5000
72
5072









Scores of the utility layers of the systems are as follows.









TABLE 25







Scores of the utility layers













System 1
System 2
System 3
















Level 1 speech data
0.747
0.716
0.7



Level 2 speech data
0.757
0.717
0.695



Level 3 speech data
0.744
0.679
0.642



Level 4 speech data
0.736
0.705
0.69



Level 5 speech data
0.678
0.6001
0.5616



Total score
3.662
3.4171
3.2896










Scores of the support layers of the systems are as follows.









TABLE 26







Scores of the support layers of the systems













System 1
System 2
System 3







Support layer
0.43
0.396
0.312










Weights of the utility layer indexes and the support layer indexes of the system 4 obtained by using the improved group-G2 method are as follows.









TABLE 27







Weights of the utility layer indexes and the support layer indexes obtained by


using the improved group-G2 method



















Minimum







Assignment
Assignment
Assignment
class non

Risk
Interval




interval
interval
interval
empty
Comprehensive
attitude
mapping
Index


Index
Di(1) of an
Di(2) of an
Di(3) of an
intersection
assignment
factor
function
weight


xi
expert 1
expert 2
expert 3
DiAJ
interval
λi
ƒλi
wi


















Index
[1, 1]
[1, 1]
[1, 1]
[1, 1]
[1, 1]
0
1
0.37


1










Index
[1.5, 2]
[1.4, 1.8]
[1.3, 2.3]
[1.5, 1.8]
[1.5, 1.8]
0.152
1.7
0.63


2

















Weights of the utility layer indexes and the support layer indexes of the system 4 obtained by using the improved CRITIC method are as follows.















Utility layer
Support layer







Information entropy
0.5264316348395962
0.6255467860676905


Information carrying
0.06278809
0.06358222


capacity




Objective weight
0.47918135
0.52081865









Comprehensive weights of the utility layer and the support layer of the system 4 are calculated as follows.









TABLE 29







Comprehensive weights of the utility layer and the support layer










Utility layer
Support layer







0.6103762499483129
0.38962375005168703










Optimal values and worst values of the scores of the utility layer indexes and the support layer indexes obtained from simulation statistics are substituted into the improved TOPSIS method to calculate a final score of the system 4.


Optimal values and worst values of the scores of the utility layer and the support layer are as follows.









TABLE 30







Optimal values and worst values of the scores of the


utility layer and the support layer










Score of the utility layer
Score of the support layer





Optimal value
5
1


Worst value
0
0









Scores of the utility layer of the system 4 under levels 1-5 speech data are calculated according to the dynamic weights of the utility layer indexes under the speech data with different quality levels.









TABLE 31







Scores of the utility layer of the system 4 under


levels 1-5 speech data














Level 1
Level 2
Level 3
Level 4
Level 5




speech
speech
speech
speech
speech
Total



data
data
data
data
data
score





Score
0.7762
0.7801
0.8161
0.7924
0.721
3.8858









According to the fixed weight of each support layer index, the score of the support layer of the system 4 is calculated as 0.8093. According to the fixed weights of the utility layer and the support layer, the final score of the system 4 is calculated as 0.79. According to the gradation principle of the recognition performance of the speech recognition system, the level of the recognition performance of the to-be-evaluated speech recognition system is determined as level 4.


In an embodiment, an evaluation device of the ATC speech recognition system adopts a Core i7-12700 processor, and a memory adopts a 980 PRO iT solid state disk of Samsung, thus making at least one processor can execute any one steps of the evaluation method.


The above is merely the embodiment of the disclosure, and it is not used to limit the disclosure. Any modification, equivalent substitution and improvement made within a spirit and principle of the disclosure should be included in a scope of protection of the disclosure.

Claims
  • 1. An evaluation method of an air traffic control (ATC) speech recognition system, comprising: S1, obtaining ATC speech data, grading, based on a quality level of the ATC speech data, the ATC speech data to obtain ATC speech data with different quality level, and inputting the ATC speech data with different quality level into a to-be-evaluated ATC speech recognition system;S2, constructing an evaluation index system; wherein indexes of the evaluation index system comprise utility layer indexes and support layer indexes;S3, determining, by using an improved group-G2 method and an improved criteria importance through intercriteria correlation (CRITIC) method, weights of the utility layer indexes and weights of the support layer indexes under the ATC speech data with different quality levels;S4, calculating, according to the weights of the utility layer indexes and the weights of the support layer indexes under the ATC speech data with different quality levels, scores of a utility layer and scores of a support layer under the ATC speech data with different quality levels by using an improved technique for order preference by similarity to ideal solution (TOPSIS) method; and adding the scores of the utility layer to obtain a comprehensive score of the utility layer, and adding the scores of the support layer to obtain a comprehensive score of the support layer; wherein the TOPSIS method is to determine Z+ and Z− of an index through a statistical method, to thereby score a recognition performance of the to-be-evaluated ATC speech recognition system; andwherein Z+ represents a set containing optimal values of evaluation indexes of the to-be-evaluated ATC speech recognition system, and Z− represents a set containing worst values of the evaluation indexes of the to-be-evaluated ATC speech recognition system;S5, determining, according to the comprehensive score of the utility layer and the comprehensive score of the support layer, a weight of the utility layer and a weight of the support layer of the to-be-evaluated ATC speech recognition system by using the improved group-G2 method and the improved CRITIC method;S6, multiplying the weight of the utility layer and the comprehensive score of the utility layer to obtain a first product, multiplying the weight of the support layer and the comprehensive score of the support layer to obtain a second product, and adding the first product and the second product to obtain a comprehensive score of the to-be-evaluated ATC speech recognition system; andS7, determining, according to the comprehensive score of the to-be-evaluated ATC speech recognition system, a level of the recognition performance of the to-be-evaluated ATC speech recognition system.
  • 2. The evaluation method of the ATC speech recognition system as claimed in claim 1, wherein in step S2, the utility layer indexes comprise accuracy indexes and a recognition rate index, the recognition rate index comprises a processing time consumption, and the processing time consumption refers to a time consumed from a moment when the to-be-evaluated ATC speech recognition system receives completely inputted ATC speech data to a moment when the to-be-evaluated ATC speech recognition system recognizes a result and outputs the result.
  • 3. The evaluation method of the ATC speech recognition system as claimed in claim 2, wherein the accuracy indexes comprise a sentence accuracy, a word accuracy and a keyword accuracy, and a calculation expression of the keyword accuracy is expressed as follows:
  • 4. The evaluation method of the ATC speech recognition system as claimed in claim 1, wherein in step S2, the support layer indexes comprise reliability indexes and maintainability indexes.
  • 5. The evaluation method of the ATC speech recognition system as claimed in claim 1, wherein in step S4, calculation formulas of Z+ and Z− of the index in the improved TOPSIS method are as follows: Z+=(max{z11,z21, . . . ,zn1},max{z12,z22, . . . ,zn2}, . . . max{z1n,z2n, . . . ,znn};Z−=(min{z11,z21, . . . ,zn1},min{z12,z22, . . . ,zn2}, . . . min{z1n,z2n, . . . ,znn};wherein z11 represents a first index value of a first system, z21 represents a first index value of a second system, z22 represents a second index value of a second system, z2n represents a n-th index value of the second system, and znn represents a n-th index value of a n-th system.
  • 6. The evaluation method of the ATC speech recognition system as claimed in claim 1, wherein in step S7, the determining, according to the comprehensive score of the to-be-evaluated ATC speech recognition system, a level of the recognition performance of the to-be-evaluated ATC speech recognition system comprises: determining, in response to the comprehensive score being in a range of 0.0-0.3 scores, containing 0.3 scores, the level of the recognition performance to be level 1;determining, in response to the comprehensive score being in a range of 0.3-0.5 scores, containing 0.5 scores, the level of the recognition performance to be level 2;determining, in response to the comprehensive score being in a range of 0.5-0.7 scores, containing 0.7 scores, the level of the recognition performance to be level 3;determining, in response to the comprehensive score being in a range of 0.7-0.9 scores, containing 0.9 scores, the level of the recognition performance to be level 4; anddetermining, in response to the comprehensive score being in a range of 0.9-1.0 scores, containing 1.0 score, the level of the recognition performance to be level 5.
  • 7. An evaluation system of the ATC speech recognition system, comprising: at least one processor, and a memory communicatively connected to the at least one processor; wherein the memory stores an instruction executed by the at least one processor, and the instruction is configured to be executed by the at least one processor, to make the at least one processor to execute the evaluation method as claimed in claim 1.
Priority Claims (1)
Number Date Country Kind
202211687813.3 Dec 2022 CN national
Foreign Referenced Citations (1)
Number Date Country
113256186 Aug 2021 CN
Non-Patent Literature Citations (4)
Entry
Machine English language translation of CN 113256186-A (Year: 2021).
Wang et al., “Research on Cognitive APP Evaluation of Visually Impaired Users,” 2020 International Conference on Computer Network, Electronic and Automation (ICCNEA), Xi'an, China, 2020, pp. 133-137, doi: 10.1109/ICCNEA50255.2020.00036 (Year: 2020).
Civil Aviation Flight University of China (Applicant), Claims for CN202211687813.3 (allowed), Dec. 28, 2022.
CNIPA, Notification to grant patent right for invention in CN202211687813.3, dated Feb. 14, 2023.