The present application claims priority from Japanese Patent Application No. 2016-177605 (filed on Sep. 12, 2016) and Japanese Patent Application No. 2017-100130 (filed on May 19, 2017), the contents of which are hereby incorporated in their entirety by reference into this specification. The present invention relates to a waveform disaggregation apparatus, a method and a program.
There have been various proposals for technology for non-intrusively estimating the state of an electrical device based on electrical current measured from a switchboard (distribution board) (Non-intrusive Load Monitoring: NILM, or Non-intrusive Appliance Load Monitoring: NIALM).
For example, Patent Literature 1 discloses an electrical device monitoring system that includes a data extraction means for extracting data related to a current and a phase of the current to a voltage for each of fundamental wave and harmonics, from measured data detected by a measuring sensor installed near a feeder entrance to a house of a customer, and a pattern recognition means for estimating operation state of an electrical device used by the house of the customer, based on data related to a current and a phase of the current to the voltage for each of fundamental wave and harmonics, obtained by the data extraction means.
As related technology that performs waveform disaggregation based on a probability model, Patent Literature 2 for example, obtains data representing a sum of electrical signals of 2 or more electrical devices including a first electrical device, and by processing the data by using a probability generating model, generates an estimated value of an operation state of the first electrical device to output the estimated value of electrical signals of the first electrical device. The probability generating model has factors that represent 3 or more states and that correspond to the first electrical device. The probability generating model is a Factorial Hidden Markov Model (FHMM). The Factorial HMM has a second factor corresponding to a second electrical device among the 2 or more electrical devices, and by processing the data using the Factorial HMM, generates a second estimated value of a second electrical signal of the second electrical device, calculates a first individual distribution of estimated value of an electrical signal of the first electrical device, uses the first individual distribution as a parameter of a factor corresponding to the first electrical device, calculates a second individual distribution of the second estimated value of the second electrical signals of the second electrical device, and uses the second individual distribution as a parameter of a factor corresponding to the second electrical device.
With a normal HMM (Hidden Markov Model), one state variable St corresponds to observed data Yt at time t, but in Factorial HMM there are multiple (M) state variables St, St(1), St(2) to St(M), and one observation data item Yt is generated based on the multiple state variables St(1) to St(M). The state variables St(1) to St(M) respectively correspond to electrical devices. State values of the state variables St(1) to St(M) correspond to states (operation state, for example, ON, OFF) of the electrical devices. In HMM an EM (Expectation-Maximization) algorithm used for estimating a parameter(s) from output (observation data) is an algorithm that maximizes logarithmic likelihood of observation data by repeating E (Expectation) and M (Maximization) steps, and includes the following steps 1 to 3.
1. Set initial parameters.
2. Compute expected value of likelihood of model based on distribution of presently estimated latent variables (E Step).
3. Find parameters to maximize expected value of likelihood obtained in the E Step (M Step). The parameters obtained in the M Step are used to determine distribution of latent variables used in a subsequent E Step, and steps 2 and 3 are repeated until the expected value converges (no longer increases).
Patent Literature 3 discloses an electrical device estimation apparatus including a data acquisition means for acquiring time series data for total value of consumption current of plural electrical devices, and a parameter estimating means for finding model parameters with operation states of the plural electrical devices being modeled by a probability model, based on the acquired time series data. The probability model is a Factorial HMM. The data acquisition means converts a total value of acquired consumption current into non-negative data, and the parameter estimating means, in parameter estimation processing by the EM algorithm, finds a parameter W(m) of observation probability as the model parameter, by maximizing a likelihood function which is a degree describing a total value pattern for the consumption current represented by the time series data, by the Factorial HMM, under a constraint condition that observation probability parameter W(m) corresponding to a current waveform pattern of factor m of the Factorial HMM, is non-negative.
Here, a description is given of an outline of waveform disaggregation using Factorial HMM disclosed in Patent Literature 2.
A state estimation section 212 performs state estimation that estimates operation state of each home electric appliance, using current waveform Yt from a data acquisition unit 211, and model parameter φ of an overall model which is the overall model of electric appliances in a household stored in a model storage section 213.
The model learning section 214 performs model learning to update the model parameter φ of the overall model stored in a model storage unit 213, using the current waveform Yt supplied from the data acquisition unit 211 and the estimation result (operation state of each home appliance) of state estimation supplied from the state estimation section 212. The model parameter φ includes initial probability, distribution, and characteristic waveform W(m).
The model learning section 214 performs waveform disaggregation learning to obtain (update) the current waveform parameter as a model parameter, using current waveform Yt supplied from the data acquisition unit 211, and operation state of each home appliance supplied from the state estimation section 212, and updates the current waveform parameter W(m) stored in the model storage unit 213, by the current waveform parameter obtained by waveform disaggregation learning.
The model learning section 214 performs disaggregation learning to obtain (update) the distribution parameter as a model parameter, using current waveform Yt supplied from the data acquisition unit 211, and operation state of each home appliance supplied from the state estimation section 212, and updates distribution parameter C stored in the model storage unit 213, by the distribution parameter obtained by distribution learning thereof.
The model learning section 214 performs state change learning to obtain (update) the initial state parameter as model parameter φ, and a state change parameter, using operation state of each home appliance supplied from the state estimation section 212, and updates each of the initial state parameter stored in the model storage unit 213 and the state change parameter, by the initial state parameter obtained by the state change learning and the state change parameter. HMM can be used as an overall model stored in the model storage unit 213. The data output section 216 obtains and displays, on a display apparatus or the like, consumption power of home electrical appliances represented by respective home electrical appliance models using the overall model stored in the model storage unit 213.
As further related technology, in Patent Literature 4, current waveform data is extracted, which is obtained by averaging total load current for one cycle of commercial power supply frequency, based on total load current and voltage measured at a prescribed position in a service wire of a customer area, and convex point information is extracted that relates to a convex point indicating a point where change in current value turns from increase to decrease, or a point of turning from decrease to increase, from the averaged current waveform data. The estimation section stores in advance an estimation model associating a type of an electrical device with convex point information and consumption power. The estimation section individually estimates consumption power of an electrical device being operated, based on convex point information extracted by the data extraction unit and estimation model.
Patent Literature 5 discloses a power estimation apparatus that receives current waveform and voltage waveform measured for an electrical device that consumes power from one or a plurality of power sources and estimates consumption power of the electrical device from the current waveform of the electrical device, includes a power estimation section that estimates electrical power for each electrical device based on data of the received current waveform and voltage waveform; a holding unit that holds power consumption patterns representing characteristics of consumption power and change amount of the consumption power, for each electrical device; and an estimation power correction unit that decides whether or not the electrical power estimated by the electrical power estimation section matches the electrical power consumption pattern held by the holding unit, and in a case where it is decided that there is no match, corrects the electrical power according to the electrical power consumption pattern.
An apparatus consumption electrical power estimation apparatus disclosed in Patent Literature 6 includes a device feature learning section, a device feature database, an operation state estimation section, and a consumption power estimation section. The device feature learning section obtains a feature value of an operation state of an apparatus from electrical current or power frequency obtained from time series data of voltage and current measured in a power supply path. The device feature database stores the obtained feature value of the operation state of the apparatus. The operation state estimation section estimates the operation state of the device based on harmonics feature values obtained from harmonics of electrical current or power, and a feature value(s) of operation state of the device stored in the device feature database. The consumption power estimation section estimates consumption power of the device based on the estimation operation state.
It is noted that for the FHMM, EM algorithm, Gibbs-Sampling and the like, Non-Patent Literature 1 for example may be referred to.
An analysis of the related technology is given below. In the above described related technology that relates to waveform disaggregation, it is not possible, for example, to perform waveform disaggregation for a plurality of units with identical or substantively identical configuration. Or, even if waveform disaggregation can be performed, accuracy may be reduced. As in a production line, for example, it is a fact that there is no example of application of waveform disaggregation to a case (system) where there are a plurality of devices of the same type.
Accordingly, the present invention was invented in consideration to the above described issues, and it is an object thereof to provide a waveform disaggregation apparatus, a method and a program, each enabling to disaggregate, from a composite signal waveform, signal waveforms between units of identical or substantively identical configuration, for example.
According to an aspect of the present invention there is provided a waveform disaggregation apparatus comprising:
a storage apparatus that stores, as a model of an operation state of a unit, a first state transition model including a segment in which each state transition occurs along a one directional single path; and
an estimation section that receives a composite signal waveform of a plurality of units including a first unit that operates based on the first state transition model,
the estimation section performing, at least based on the first state transition model, estimation of a signal waveform of the first unit from the composite signal waveform to separate the signal waveform therefrom.
According to an aspect of the present invention there is provided a computer-based waveform disaggregation method comprising:
regarding a composite signal waveform of a plurality of units including a first unit that operates based on a first state transition model, the first state transition model including a segment in which each state transition occurs along a one directional single path,
performing, based on the first state transition model, estimation of a signal waveform of the first unit from the composite signal waveform to separate the signal waveform therefrom.
According to an aspect of the present invention there is provided a program that causes a computer to execute processing comprising:
receiving a composite signal waveform of a plurality of units including a first unit that operates based on a first state transition model, the first state transition model including a segment in which each state transition occurs along a one directional single path; and
performing, based on the first state transition model, estimation of a signal waveform of the first unit from the composite signal waveform to separate the signal waveform therefrom. According to the present invention, there is provided a computer readable storage medium that stores the above described program (for example, a non-transitory computer readable recording medium such as semiconductor storage such as RAM (Random Access Memory), ROM (Read Only Memory), EEPROM (Electrically Erasable and Programmable ROM) or the like, a HDD (Hard Disk Drive), CD (Compact Disc), DVD (Digital Versatile Disc) or the like).
According to another aspect of the present invention, the waveform disaggregation apparatus may be configured to include an estimation section that estimates and disaggregates a signal waveform of a plurality of units from a composite signal waveform of the plurality of units, and an anomaly estimation section that receives a signal waveform disaggregated for each unit by the estimation section, calculates anomaly level indicating a degree of anomaly, from the signal waveform or a prescribed state to detects an anomaly of the unit.
According to the present invention, it is possible, for example, to separate a signal waveform between units having identical or substantively identical configurations, from a composite signal waveform.
The following describes one of modes of the present invention.
According to the embodiment of the present invention, as illustrated schematically in a model 121 of
According to an example embodiment of the present invention, the plurality of units include a second unit identical or of identical type as the first unit, and the estimation section 11 may be configured to disaggregate a composite signal waveform of the first and second units into a signal waveform of the first unit and a signal waveform of the second unit, based on the first state transition model of the first unit and a state transition model of the second unit.
According to an example embodiment of the present invention the first and second units may be any of:
first and second units provided within one equipment configuring one production line,
first and second facilities configuring one production line, and
a first unit of a first equipment configuring a first production line and a second unit of a second equipment configuring a second production line. Alternatively, the first and second unit may be first and second personal computers (PCs) of identical or substantively identical configuration (first and second home electrical appliances).
According to an example embodiment of the present invention, a signal, a waveform of which is subjected to disaggregation may be electrical current, voltage, power or the like.
According to an example embodiment of the present invention, it is possible to disaggregate waveforms of the first unit and the second unit from a composite waveform of the plurality of units including at least the first unit with operational constraint imposed thereto and a second unit identical or of a substantively identical configuration to the first unit.
Next, referring to
In
1-1 represents a waveform (holding a constant level) of a stop state (state (1));
1-2 represents a waveform of a certain work operation (state (2)); and
1-3 represents a waveform of another work operation (state (3)). It is noted that in the respective waveforms 1-1 to 1-3 of
Here, constraint I and constraint II are imposed on factor 1. However, only one of either constraint I and constraint II may be imposed.
Constraint I: when in state (2) at a certain time t, at a next time t+1, in state (3).
Constraint II: when in state (2) at a certain time t, at a previous time t−1, in state (1).
As an example of constraint II, as illustrated in
It is noted that correspondence of combination (3×3) of states (1) to (3) of factor 1 and factor 2 and composite waveforms is as schematically illustrated in
In
Similarly, at time t=4, it is understood that there exists a combination of states (1) and (3) as states of factor 1 and 2. However, it is not known which of factor 1 and factor 2 is in state (1) and which is in state (3).
On the other hand, as in an example embodiment of the present invention, in a case of a constraint on state transition, as illustrated in
Referring to
Due to constraint I on factor 1, since state (3) exists at time after state (2), it is confirmed that factor 1 at time t=4 is in state (3). Therefore, factor 2 at time t=4 is in state (1). It is noted that correspondence between composite waveforms and factor 1 and 2 states, shown schematically in
In this way, according to an example embodiment of the present invention, by introducing a constraint to a state transition, it is possible to confirm a state of each of units which have identical configuration.
Using the above described constraints is advantageous with regard to amount of computation, description of which is given later.
Above a description has been given of configuration and operational principles of an example embodiment of the present invention. Below, a description is given of several example embodiments.
Referring to
A current sensor 102 measures power supply current (composite power supply current of respective facilities of the production line) of, for example, the main flow of a distribution board 103. The current sensor 102 transmits measured current waveform (digital signal waveform) via a communication apparatus 101 to a waveform disaggregation apparatus 10. The current sensor 102 may be configured by a CT (Current Transformer) (for example, Zero-phase-sequence Current Transformer: ZCT)) or a Hall element. The current sensor 102 may perform sampling of current waveform (analog signal) by an analog digital transformer which is not illustrated, and transform the sampled signal to a digital signal waveform, and perform compression coding by an encoder which is not illustrated, to perform wireless transmission of the compression coded data to the communication apparatus 101 by a W-SUN (Wireless Smart Utility Network) or the like.
It is noted that the communication apparatus 101 may be arranged in a factory (building). The waveform disaggregation apparatus 10 may be arranged inside a factory or may be implemented on a cloud server connected with the communication apparatus 101 via a wide area network such as the Internet.
The storage apparatus 12 stores state transition models that model transitions of operation states for respective devices (for example, loader 105, unloader 111, solder printer 106, inspection machines 1, 2 (107, 110), mounters 108A to 108C, reflow oven 109) that configure the line of
It is noted that in the first example embodiment, where an equipment has identical plural units, in order to perform waveform disaggregation thereof, a state transition model of at least one unit (first unit) includes a model corresponding to a state transition diagram including a one-directional single path segment.
An estimation section 11 estimates and performs estimation and disaggregation of respective power supply current waveforms of respective units, based on a state transition model stored in the storage apparatus 12, with respect to a composite power supply current obtained by the current waveform acquisition section 13.
It is noted that in
In model 123 of the first unit, for a one-directional single path segment (state p1(1) to p3(1)), the state of the first unit corresponds to an operation constraint of the first unit that when the state (hidden state St(1)) at a time t is p1(1), the state (hidden state St+1(1)) at a next time t+1, is p2(1) with transition probability=1. It is noted that (1) on a shoulder of operation state p1(1) represents factor 1, notation of which corresponds to (1) on a shoulder of state variable St(1) and (2) on a shoulder of operation state p1(2) of the model 124 of the second unit represents factor 2, notation of which corresponds to (2) in the shoulder of state variable St(2).
An output section 14 outputs current waveforms of respective units for which estimation and disaggregation have been performed by an estimation section 11 (
In the first example embodiment, a unit which is a target for estimation and disaggregation of current waveform and on which an operation constraint is imposed (state transition model includes one directional single path segment), may, in a case where an equipment (e.g., a mounter) of
State transition probability P(St|St−1) between states is given as below.
P(St=pk|St−1=pk−1)=P(St=w|St−1=pT)=1 (1)
P(St=p1|St−1=w)=α (2)
P(St=w|St−1=w)=1−α (3))
The above equation (1) indicates that when a value (operation state) of state variable St−1 at time t−1 is pk−1, a probability that a value (operation state) of state variable St at subsequent time t transitions to pk is 1 (k=1 to T), and when a value (operation state) of state variable St−1 at time t−1 is pT, a probability that a value (operation state) of state variable St at subsequent time t transitions to W is 1.
The above equation (2) indicates that when a value (operation state) of state variable St−1 at time t−1 is w (waiting state), a probability that a value (operation state) of state variable St at subsequent time t, transitions to p1, is α (0<α<1).
The above equation (3) indicates that when a value (operation state) of state variable St−1 at time t−1 is w (waiting state), a probability that a value (operation state) of state variable St at subsequent time t transitions to w, is 1−α.
In the first example embodiment, in estimating and learning of current waveform parameters of a unit (factor) using an operation state model (state transition model) of a unit stored in the storage apparatus 12, it is, as a matter of course, possible to use, as disclosed in Non-Patent Literature 1, an EM algorithm, Gibbs sampling, Completely Factorized Variational Inference, Structured Variational inference or the like. Among these, Patent Literature 3 describes an example of estimation processing of current waveform parameters and the like using Completely Factorized Variational Inference, Structured Variational Inference. In Patent Literature 3, Structured Variational Inference is described as an example of E step, and in M step corresponding to this, Completely Factorized Variational Inference is used. It is noted that in the first example embodiment for example, Structured Variational Inference may be used (refer to Non-Patent Literature 1), though not limited thereto.
In Structured Variational Inference, as described in Appendix D of Non-Patent Literature 1, a parameter ht(m) that minimizes Kullback-Leibler divergence) KL which is a similarity measure of probability distribution, may be derived as below. It is noted that with Structured Variational Inference of Non-Patent Literature 1, Kullback-Leibler divergence KL is given below.
Z in the equation (4) is a normalized constant for posterior probability sum being 1 when an observation sequence is given, and ZQ is a normalized constant of probability distribution (expression (C.1), (C.3) of Appendix C of Non-Patent Literature 1. It is noted that H({St, Yt}), HQ({St}) are defined in expressions (C.2), (C.4) of Appendix C).
With partial derivative of the above equation (4) with log hσ(m), the following expression (5) is given.
With regard to ht(m) that minimizes Kullback-Leibler divergence KL, by having content in the parentheses [ ] of above equation (5) as 0, the following equation (6a) is obtained. Note that equations (6a) and (6b) are obtained for m=1 to M (number of factors).
Where, Δ(m)=diagonal (W(m)′C−1W(m)) (diagonal indicates diagonal component of matrix).
Residual ˜Yt(m) is defined as below.
Parameter ht(m) is an observation probability related to state variable St(m) in Hidden Markov Model m. Using a forward and backward algorithm using this observation probability and state transition probability matrix A1, j(m), a new set of expected value <St(m)> is obtained, and feedback is made to equations (6a) and (6b).
In the example of
With regard to state estimation of respective times, a parameter j that can best explain observation data X(Yt) is obtained (maximum likelihood estimation).
It is noted that the expression (7) may be given as below when notation is matched to Non-Patent Literature 1.
Here, supplementing the denotation, with regard to Non-Patent Literature 1, St(m) used in the description of
S
t
(m)
is represented by a vector known as “1-of-N representation” (refer to Non-Patent Literature 2). For M states, a vector of “1-of-M representation” representing state j becomes a vector in which only element j is 1 and the remainder are 0. Taking an expected value of this vector, respective elements form a vector, each element representing a probability of taking each state.
S
t,j
(m)
=P(St,j(m)=1|X)
Here, a right side of the above equation
P(St,j(m)=1|X)
corresponds to
P(St(m)=j|X)
of the expression (7). That is, with regard to St,j(m), the following holds. (Probability that St,j(m) is 1)=(Probability that state of factor m at time t is j)
Next, as a specific example of the first example embodiment, in the production line of
Here, a defined operation constraint is imposed on the first half unit, though not limited thereto.
In
An operation constraint as in the first half unit (stage 1) need not be imposed on an operation of the latter half unit (stage 2). Alternatively, an operation constraint similar to the stage 1 may, as a matter of course, be imposed on an operation of the latter half unit (stage 2). It is noted that the stages 1 and 2 may each be configured to operate independently, or they may operate in synchronization.
In
In
It is noted that in a case where an operation constraint similar to the first half unit (stage 1) is imposed on an operation of the latter half unit (stage 2), it is possible to obtain a current waveform of the latter half unit (stage 2), similarly to the first half unit.
From
In
It is noted that in a case of performing waveform disaggregation machine learning for each unit (factor) in the estimation section 11 of
In
In
Therefore, MAE represents error expressing how much each cycle time of each individual product is deviated.
The first example embodiment illustrates an example of application to a technique enabling visualization of operation state of a plurality of production facilities using a single sensor, for example.
As described above, the first example embodiment is effective for improving production line efficiency.
In the first example embodiment, by applying Factorial HMM where respective factors represent cycle operations of facilities, to core current waveform data, visualization of product flow in a production line by a single sensor is made possible.
In estimating cycle time, as illustrated in
According to the first example embodiment, by imposing an operation constraint on at least one unit (for example, first half unit (stage 1)), among units with identical or almost identical configuration (having a one directional single path segment in a state transition model), it is possible, for example, to disaggregate current waveforms among units with identical or almost identical configuration, from a composite current waveform of a plurality of units.
In a second example embodiment, as illustrated in
The model creation section 15 may have a configuration provided with a parameter learning function. The parameter learning function fixes a defined operation constraint imposed on a unit (transition state model having a one directional single path segment), and finds a solution of a parameter optimization problem, based on output of an estimation section 11, from observation data (for example, composite current waveform). A parameter to be optimized, may be a transition probability of a state transition model of a unit where a defined operation constraint is imposed.
Alternatively, the model creation section 15 may include a model structure learning function. The model structure learning function sequentially changes, for example, from an initial setting value, a structure of a fixed operation constraint (transition state model having a one directional single path segment) imposed on a unit to find a solution of an optimization problem. As the structure of the defined operation constraint to be changed, an issue may be on which state transitions, several constraints (one directional single path segment) are imposed. The fixed operation constraint(s) imposed on a unit may be changed and based on a result of estimation disaggregation of waveform by the estimation section 11 based on observation data, an operation constraint providing optimum waveform disaggregation may be determined. Models 125 and 126 of a plurality of units (unit m, and unit n: where m and n are prescribed positive integers that are different from each other) of the storage apparatus 12 illustrate state transition models of respective units created by the model creation section 15. In the model 125, state pm1-pm3 form a one directional single path segment corresponding to operation constraints of the unit m. It is noted that, similar to the first example embodiment, a model formed by combination of state transitional models of this plurality of units clearly may configure a Factorial HMM model.
According to the second example embodiment, model creation may be made automatic, and by parameter optimization and model learning, it is possible to improve model accuracy and to set suitable operation constraints.
In the waveform disaggregation apparatuses 10 and 10A, output from an output section 14 may be a state string (operation state: p1 to pT in
Input of waveform disaggregation apparatuses 10 and 10A may be waveform, frequency component, principal component, root mean square value, average value, power factor or the like of voltage or current. In a case other than where output is power (operation state), a signal acquisition unit that obtains input (acoustic signal, oscillation, communication amount, etc.) other than power may be provided.
In the first and second example embodiments, mainly, the application to a production line facility is described as an example, but the example embodiments of the present invention is not limited to production line facility and may be applied to domestic or enterprise personal computers (PC) or the like.
The following describes a third example embodiment of the invention. In the third example embodiment, a plurality of identical personal computers are connected to a distribution board, a printer or the like is additionally connected, and waveforms for individual devices are disaggregated in a case where a plurality of identical personal computers are connected. For example, a power supply current (a composite current waveform of electrical home appliances including personal computers 24A and 24B, and a printer 25 that are connected via a branch breaker to the distribution board 22) which is detected by a current sensor 23 that detects a current flowing in a main line (or branch breaker) of a distribution board 22 in
An operation state of a personal computer after power up, generally depends on how a user uses the personal computer. Thus, imposing a fixed operational constraint may be almost impossible.
However, a transition of an operation state of a personal computer power supply ON (at powering up) operation or a power supply OFF (at shutting down), operation is basically in a one directional single path transition. For example, in a case where types (model, machine type, etc.) are identical, or where OSs (Operating Systems) are identical, or a case where applications that start up automatically after the OS starts up, or applications that operate automatically before shutdown are identical, a power-up sequence or a shutdown sequence for the personal computer in question are basically identical (excepting where start up does not happen due to some trouble). Alternatively, a model may be created by a model creation section (15 in
As illustrated in
According to the third example embodiment, it is possible to extract a waveform of an individual personal computer on which a fixed operational constraint is imposed, from a composite current waveform of a plurality of identical personal computers, for example. As a result, it is possible to estimate an operation state (what time the power supply is turned ON or OFF, etc.) of the identical personal computers.
As described above, in the above described respective example embodiments it is possible to disaggregate waveforms of a plurality of units with identical configuration by including a one directional single path segment in a model (state transition model) of an operation state of a unit. That is, it is possible to distinguish which unit corresponds to which waveform. In addition, computation amount (quantity) is reduced by including a one directional single path segment in the state transition model. A description is given below concerning this point.
In a forward algorithm and a backward algorithm used in state estimation, multiplication of a transition probability matrix and a probability vector is necessary. Since a transition probability matrix A is a sparse matrix (many elements of the matrix are 0), when calculating a product of the transition probability matrix A and the probability vector P, it is possible to greatly reduce computation amount by excluding zero elements from the computation in advance.
Similarly, in the Viterbi algorithm used in estimating a state, computation is necessary to obtain a maximum value in each column of a product of elements of the transition probability matrix and elements of the probability matrix. In this case also, by removing zero elements of the probability matrix from computation of the maximum value in advance, it is possible to greatly reduce computation amount.
When a constraint as in
When a probability that a value of a state variable St(1) of factor 1 is a state #i, and a value of a state variable St(2) of factor 2 is state #j, is given at a certain time t,
αi,j=P[St(1)=i,St(2)=j] (11)
a probability that a value of a state variable St+1(1) of factor 1 at a next time t+1 is state #k, and a value of a state variable St+1(2) of factor 2 is a state #1, is given by the following expression (12).
Here, the Kronecker product
A⊗B
with A=(aij) being an m×n matrix, B=(bk1) being a p×q matrix, is a mp×nq partition segmented matrix.
For example, for the transition probability matrix A (3×3) of
In the above matrix, there are 54 non-zero elements among 9×9=81 matrix elements. In computation of a product of this matrix and a vector product in a forward algorithm or a backward algorithm, or computation of a maximum value appearing in a Viterbi algorithm, it is possible to reduce computation amount by skipping calculation of zero elements. When the number of operation constraints according to the present example embodiment increase, non-zero elements become fewer and computation time is shortened.
Next a description is given concerning computation amount in iterations of E step in Structured Variational Inference according to the present example embodiment.
A computation amount for a product of a matrix and a vector is proportional to the number of non-zero elements in the matrix (the above expression 9). In a normal Factorial HMM with a non-sparse matrix, there are M{circumflex over ( )}2 non-zero elements for M states in a transition probability matrix ({circumflex over ( )} is exponential operator).
In the present example embodiment, as illustrated in the example of
Next, with related technology (Patent Literature 2) described with reference to
In the related technology (Patent Literature 2), in order that elements of a transition probability matrix be zero by chance, as a result of learning, in M step, in an updating expression of a state transition probability matrix Ai, j(m) (in expression (15) of Patent Literature 2, Ai,j(m)new is pi,j(m)new).
a right side must be zero.
<St−1,i(m), St,j(m)> is an element of i-th row and j-th column of the K×K posterior probability <St−1(m)St(m)>, and represents a state probability of a state being in state #j at a next time t, when the state is in state #i at time t−1. <St−1,i(m)> represents a state probability of a state being in state #i at time t−1.
In M step, a model learning section 214 of
In order that a numerator of the right side of the above expression (15) is zero, for posterior probabilities <St−1(m) St(m)′> (expression (11) of Patent Literature 2),
a sum of numerators on the right side must all be zero. It is noted that P(z|w) is a probability of a transition to a combination z of states from a combination w of states. This is obtained as a product of as from P(1)i(1),j(1) which is a transition probability from a state #i(1) of factor #1 configuring a combination w of states to a state #j(1) of factor #1 configuring a combination z of states, to P(M)i(M),i(M) which is a transition probability from a state #i(M) of factor #M configuring a combination w of states to a state #j(M) of factor #M configuring a combination z of states. The transition probability P(St|St−1) is given by the following expression (17).
With respect to factor m, P(St(m)|St−1(m)) is a probability of transitioning to state St(m) at time t, when being in state St−1(m) at time t−1.
An observed probability P(Yt|St) is given by the following (expression (4) of Patent Literature 2).
P(Yt|St)=|C|−1/2(2π)−D/2 exp {−½(Yt−μt)C−1(Yt−μt)} (18)
A dash (′) represents a transpose. From the above expression, P(Yt|z)>0.
Since a forward probability αt−1,w of Factorial HMM and backward probability βt, z of Factorial HMM are probability variables, a certain w and z exist, and
αt−1,w>0, βt,z>0. (19)
Therefore, in order that “elements of a transition probability matrix after update are zero”, “elements of the transition probability matrix before update are zero”.
That is, as long as elements of the transition probability matrix are not made to zero before learning, they are not zero after learning. From the above, it has been shown that a constraint introduced in an example embodiment of the present invention is not something that can be automatically learned by a known learning algorithm such as an EM algorithm or the like.
Next, a description is given regarding a fifth example embodiment of the invention, with reference to
The anomaly estimation section 16 of the waveform disaggregation apparatus 10B of the fifth example embodiment receives a signal waveform disaggregated by the estimation section 11 that estimates and disaggregates signal waveforms of a plurality of individual units, based on a state transition model, from a composite signal waveform, and detects an anomaly in a unit from the disaggregated signal waveform or a prescribed state. The state transition model, as a model of operation states of a unit, may preferably have a configuration including a first state transition model having a segment for transition along one directional single path.
In related technology, in a case of performing anomaly monitoring of a system using a waveform of electrical current or the like, when the system includes a plurality of units, it is not easy to detect in which unit an anomaly occurs.
The reason for this is as follows. When performing anomaly monitoring using signal waveforms of individual units, a large number of sensors are required for each individual unit, and as a result, cost increases (rises). Instead of installing sensors in individual units, in a case of performing anomaly monitoring using an entire waveform (composite signal waveform) of a system including a plurality of units, it may be possible to detect an occurrence of an anomaly from the entire waveform of the system, but it may not be easy to detect in which unit the anomaly occurs.
According to the fifth example embodiment, in a system including a plurality of units, by performing waveform disaggregation of an entire waveform (composite signal waveform of a plurality of units) of the system measured by a small number of sensors, with high accuracy, for each unit, it is possible to detect in which unit an anomaly occurs.
For example, regarding a plurality of units of identical or nearly identical configuration, even when accuracy in disaggregation accuracy in which a composite waveform is disaggregated into waveforms of the individual units in related technology, it is possible to detect a unit in which an anomaly occurs with good accuracy, according to the fifth example embodiment.
While there is no particular limitation, for example, in a case of a facility where a plurality of units configure a production line, by monitoring for “a situation (anomaly) which is different from normal”, it is possible to detect and cope with a failure of the facility or quality anomaly of products at an early stage, as a result of which it is possible to reduce production stoppage time (down time) and to improve production yield.
As another example, in a case where a plurality of units includes personal computers, by monitoring for a situation which is different from normal, it is possible to detect and cope with, at an early stage, contamination by malware (unauthorized software) in a personal computer, for example. As a result, it is possible to reduce risks with regard to information security.
In a case of the above described examples, situations often occur where a plurality of units (production facilities, personal computers, or the like) have identical or nearly identical configurations. In such a case, it is not easy to detect in which unit and in which operation an anomaly occurs, only with simple monitoring for a situation which is different from normal.
According to the fifth example embodiment, for example, even in a case where there are a plurality of units of identical or nearly identical configurations, it is possible to detect in which unit and in which operation an anomaly occurs.
The anomaly detection section 161 calculates anomaly level indicating an occurrence degree of anomaly for a waveform disaggregated for each unit, based on a disaggregating result of a signal waveform by an estimation section 11, and by comparing the anomaly level with a predetermined threshold, for example, decides whether or not there is an anomaly.
In the anomaly detection section 161, as an example of anomaly level, for example, KL divergence at each point of time may be used. KL divergence at each point of time corresponds to an extraction of contribution at time tin expression (4), and may be obtained by the following expression.
Here, as for values of variables <St(m)> and ht(m), values are used that have been estimated, for example, by the estimation section 11, as described in the first and second example embodiment. In this case, the KL divergence at each point of time indicates a measure of difference between model distribution and measured value Yt, and it may be considered that the more an anomaly is included in the measured value, the greater a value of KL divergence.
Therefore, in the anomaly detection section 161, it is possible to detect an occurrence of an anomaly according to whether or not a value KLt of KL divergence at each point of time is greater than a predetermined threshold (first threshold). That is, the anomaly detection section 161 decides that an anomaly occurs in a case where KLt is greater than the first threshold.
As another example of anomaly level, for example, a marginal likelihood at each point of time may be used. A marginal likelihood at each point of time is a probability density where a measured value Yt at time t is obtained from a model. A marginal likelihood Lt at each point of time is obtained by the following expression (21) by using residual ˜Yt(m) obtained according to the expression (6b), for example.
In this case, it is considered that the more an anomaly is included in a measured value Yt, the smaller the value of marginal likelihood Lt at each point of time. Therefore, in the anomaly detection section 161, it is possible to detect an occurrence of an anomaly according to whether or not the marginal likelihood Lt at each point of time is smaller than a predetermined threshold (second threshold). That is, the anomaly detection section 161 decides that an anomaly occurs when Lt is smaller than the second threshold.
Next, an estimation is made as to in which unit (factor) an anomaly occurs, by the anomaly location estimation section 162 of the anomaly estimation section 16.
When an anomaly is detected at time t by the anomaly detection section 161, each factor m is in a state St(m). Therefore, in the anomaly location estimation section 162, by estimating a pair (m, St(m)) of the state St(m) corresponding to a factor m in which an anomaly occurs, it is possible to estimate in which unit an anomaly occurs, and in which operation of the unit the anomaly occurs.
Here, as an estimated value of a state St(m) corresponding to each factor m, it is possible to use, for example, a value of the expression (7) which is used in the estimation section 11.
By so doing, the anomaly location estimation section 162 can obtain M items of candidates, that is, candidates of pairs (m, St(m)), where m=1, . . . , M, for a factor m and a state St(m) in which an anomaly occurs.
Next, in the anomaly location estimation section 162, among M candidates of set (m, St(m)) of factor and state in which an anomaly occurs, a priority is assigned according to a value of state St(m).
The anomaly location estimation section 162 outputs the set (m, St(m)) of a factor and state that have higher priority assigned.
It is noted that in the anomaly location estimation section 162 may adopt a criterion(s) with which the priority is determined, one or a plurality of combinations of criterions below may be used (but not limited thereto).
(a) State St(m) is an internal part of a fixed constraint segment in the model 123 (
(b) Norm of weighting vector Wj(m) corresponding to state St(m)=j has a larger value.
(c) State St(m) is a state when a specific time Δt has elapsed from a start point of a segment of the fixed operation constraint, in an internal part of the fixed operation constraint segment in the model 123 (
Here, criterion (a) means that unit m is in the middle of performing repeated operations. Therefore, in the anomaly location estimation section 162, by using criterion (a), and by reflecting a general situation that “an anomaly occurs more easily in a unit in operation than in a unit that is stopped”, it is possible to correctly estimate a factor in which an anomaly occurs.
Criteria (b) means that a dimension of waveform (for example, amplitude or root mean square value of the waveform) disaggregated by the estimation section 11 is larger in unit m. For example, in a case where an input signal to the waveform disaggregation apparatus 10B is power, acoustic signal, oscillation, communication amount or the like, in general a larger signal is generated for a unit in operation in comparison with a unit that is stopped. Therefore, in the anomaly location estimation section 162, by using criterion (b), and by reflecting the situation that “an anomaly occurs more easily in a unit in operation than in a unit that is stopped”, it is possible to correctly estimate a factor in which an anomaly occurs.
Criteria (c) means that unit m which is in the middle of repeated operations, performs a specific operation. Therefore, in the anomaly location estimation section 162, by using criterion (c) and by reflecting a situation that “an anomaly occurs more easily in a unit that is in the middle of performing a specific operation than in a unit that is not in the middle of performing a specific operation”, it is possible to correctly estimate a factor in which an anomaly occurs.
In the above described example, the anomaly location estimation section 162, regarding a set of (m, St(m)) of a factor and state in which an anomaly occurs, outputs plural sets with higher priority. The anomaly location estimation section 162 may output, as another output form,
In the above described example, the anomaly location estimation section 162, determines, as a candidate of set (m, St(m)) of a factor and state in which an anomaly occurs, only one state St(m) corresponding to each factor m using the expression (7), but it is possible to use a plurality of values, as state St(m) corresponding to each factor m.
In this case, since the probability that the state of factor m is St(m)=j is <St,j(m)>, when determining a priority of a set (m, St(m)) of a factor and state in which an anomaly occurs, the anomaly location estimation section 162 may set a new criterion:
(d) a probability <St(m)=j> corresponding to state St(m)=j has a larger value. With combination of criterion (d) with the above criterion (a)-(c), the priority may be determined. In this way, for example, even in a case where a state occurs in which accuracy of waveform disaggregation deteriorates in the estimation section 11, and one state for each factor is not determined, the anomaly location estimation section 162 can output potential candidates for anomaly occurrence location.
In the fifth example embodiment, operation of the waveform disaggregation apparatus 10B may sequentially be executed (online processing) each time a waveform is obtained by the current waveform acquisition section 13. Alternatively, operation of the waveform disaggregation apparatus 10B may be executed collectively (batch processing) after a plurality of waveforms obtained by the current waveform acquisition section 13 are stored.
Here, in a case where it is necessary to shorten time from occurrence of anomaly to detection thereof, it is desirable to execute online processing to reduce holding time of a waveform. On the other hand, in a case where accuracy rather than speed of anomaly estimation is required, it is desirable to perform batch processing.
As described above, according to the fifth example embodiment, it is possible not only to perform disaggregation of a waveform of a unit, but also to detect an anomaly that occurs in a unit, and to estimate a unit in which an anomaly occurs.
It is noted that the respective disclosures of the above described Patent Literature 1-6 and Non-Patent Literature 1 and 2 are incorporated herein by reference thereto. Modifications and adjustments of example embodiments and examples may be made within the bounds of the entire disclosure (including the scope of the claims) of the present invention, and also based on fundamental technological concepts thereof. Furthermore, various combinations and selections of various disclosed elements (including respective elements of the respective appendices, respective elements of the respective example embodiments, respective elements of the respective drawings, and the like) are possible within the scope of the claims of the present invention. That is, the present invention clearly includes every type of transformation and modification that a person skilled in the art can realize according to the entire disclosure including the scope of the claims and to technological concepts thereof.
The above described example embodiments may also be described as follows (but not limited thereto).
A waveform disaggregation apparatus comprising:
a storage apparatus that stores, as a model of an operation state of a unit, a first state transition model including a segment in which each state transition occurs along a one directional single path; and
an estimation section that receives a composite signal waveform of a plurality of units including a first unit that operates based on the first state transition model,
the estimation section performing, at least based on the first state transition model, estimation of a signal waveform of the first unit from the composite signal waveform to separate the signal waveform therefrom.
The waveform disaggregation apparatus according to supplementary note 1, wherein the plurality of units include a second unit, identical to or a type thereof being identical to, the first unit, wherein the estimation section disaggregates, from a composite signal waveform of the first unit and the second unit, a signal waveform of the first unit and a signal waveform of the second unit, based on the first state transition model corresponding to the first unit and a state transition model of the second unit.
The waveform disaggregation apparatus according to supplementary note 1 or 2, wherein the first unit operating under a constraint corresponding to the segment of the first state transition model, when in a first state at a certain time, transitions, at a subsequent time, to a second state with transition probability of 1.
The waveform disaggregation apparatus according to supplementary note 2, wherein the first units the and second units comprise any out of:
first and second units within one facility, the facility configuring one production line;
first and second facilities, each configuring one production line;
a first unit of a first facility configuring a first production line, and a second unit of a second facility configuring a second production line; and
first and second home electrical appliances.
The waveform disaggregation apparatus according to any one of supplementary notes 1-4, comprising
a current waveform acquisition section that obtains a composite current waveform of the plurality of units, as the composite signal waveform.
The waveform disaggregation apparatus according to any one of supplementary notes 1-5, further including
a model creation section that creates a model of an operation state of the unit to store the model in the storage apparatus.
The waveform disaggregation apparatus according to any one of supplementary notes 1-6, wherein one state before or one state after is estimated, based on the first state transition model and a prescribed state.
The waveform disaggregation apparatus according to any one of supplementary notes 1-6, wherein the estimation section estimates a prescribed state, based on the first state transition model and a state at a preceding time or at a succeeding time.
The waveform disaggregation apparatus according to any one of supplementary notes 1-8, wherein a model of an operation state of the unit corresponds to a factor of a Factorial Hidden Markov Model (FHMM).
A computer-based waveform disaggregation method comprising:
regarding a composite signal waveform of a plurality of units including a first unit that operates based on a first state transition model, the first state transition model including a segment in which each state transition in occurs along a one directional single path,
performing, based on the first state transition model, estimation of a signal waveform of the first unit from the composite signal waveform to separate the signal waveform therefrom.
The waveform disaggregation method according to supplementary note 10, wherein the plurality of units include a second unit, identical to or a type thereof being identical to, the first unit, wherein the method comprises
disaggregating a composite signal waveform of the first unit and the second unit into a signal waveform of the first unit and a signal waveform of the second unit, based on the first state transition model corresponding to the first unit and a state transition model of the second unit.
The waveform disaggregation method according to supplementary note 10 or 11, wherein the first unit operating under a constraint corresponding to the segment of the first state transition model, when in a first state at a certain time, transitions, at a subsequent time, to a second state with transition probability of 1.
The waveform disaggregation method according to supplementary note 11, wherein the first units the and second units include any out of:
first and second units within one facility, the facility configuring one production line;
first and second facilities, each configuring one production line;
a first unit of a first facility configuring a first production line, and a second unit of a second facility configuring a second production line; and
first and second home electrical appliances.
The waveform disaggregation method according to any one of supplementary notes 10-13, comprising
a current waveform acquisition step that obtains a composite current waveform of the plurality of units, as the composite signal waveform.
The waveform disaggregation method according to any one of supplementary notes 10-15, further comprising
a model creation step that creates a model of an operation state of the unit.
The waveform disaggregation method according to any one of supplementary notes 10-15, comprising
estimating a state at a preceding time or at a succeeding time, based on the first state transition model and a prescribed state.
The waveform disaggregation method according to any one of supplementary notes 10-15, comprising
estimating a prescribed state, based on the first state transition model and a state at a preceding time or at a succeeding time.
The waveform disaggregation method according to any one of supplementary notes 10-15, wherein a model of an operation state of the unit corresponds to a factor of a Factorial Hidden Markov Model (FHMM).
A program causing a computer to execute processing comprising:
receiving a composite signal waveform of a plurality of units including a first unit that operates based on a first state transition model, the first state transition model including a segment in which each state transition in occurs along a one directional single path; and
performing, based on the first state transition model, estimation of a signal waveform of the first unit from the composite signal waveform to separate the signal waveform therefrom.
The program according to supplementary note 19, wherein the plurality of units include a second unit, identical to or a type thereof being identical to, the first unit, wherein the estimating processing comprises
disaggregating a composite signal waveform of the first unit and the second unit into a signal waveform of the first unit and a signal waveform of the second unit, based on the first state transition model corresponding to the first unit and a state transition model of the second unit.
The program according to supplementary note 19 or 20, wherein the first unit operating under a constraint corresponding to the segment of the first state transition model, when in a first state at a certain time, transitions, at a subsequent time, to a second state with transition probability of 1.
The program according to supplementary note 11, wherein the first units the and second units include any out of:
first and second units within one facility, the facility configuring one production line;
first and second facilities, each configuring one production line;
a first unit of a first facility configuring a first production line, and a second unit of a second facility configuring a second production line; and
first and second home electrical appliances.
The program according to any one of supplementary notes 19-22, comprising a current waveform acquisition processing that obtains a composite current waveform of the plurality of units, as the composite signal waveform.
The program according to any one of supplementary notes 19-23, comprising a current waveform acquisition processing that obtains a composite current waveform of the plurality of units, as the composite signal waveform.
The program according to any one of supplementary notes 19-24, comprising
estimating a prescribed state, based on the first state transition model and a state at a preceding time or at a succeeding time.
The program according to any one of supplementary notes 19-24, comprising estimating a prescribed state from the first state transition model, and one state before or one state after.
The program according to any one of supplementary notes 19-24, wherein a model of an operation state of the unit corresponds to a factor of a Factorial Hidden Markov Model (FHMM).
The waveform disaggregation apparatus according to any one of supplementary notes 1-9, further comprising
an anomaly estimation section that detects an anomaly of the unit, from the signal waveform disaggregated by the estimation section or a prescribed state.
The waveform disaggregation apparatus according to supplementary note 28, wherein the anomaly estimation section calculates anomaly level indicating an occurrence degree of anomaly, based on the signal waveform disaggregated by the estimation section or a prescribed state and compares the anomaly level with a threshold to decide whether or not an anomaly occurs.
The waveform disaggregation apparatus according to supplementary note 28 or 29, wherein the anomaly estimation section estimates either one or both of a factor in which an anomaly occurs or a state in which an anomaly occurs, based on the signal waveform disaggregated by the estimation section or a prescribed state and compares the anomaly level with a threshold to decide whether or not anomaly occurs.
The waveform disaggregation apparatus according to supplementary note 30, wherein the anomaly estimation section determines priority for a set of the factor and the state, in accordance with an estimated value of a state corresponding to a time at which the anomaly is detected, and
estimates a set of the factor and the state with the priority being high, as either one or both of a factor in which the anomaly occurs and a state in which an anomaly occurs.
The waveform disaggregation apparatus according to supplementary note 31, wherein the anomaly estimation section adopts as criterion for determining the priority, at least one of the followings:
(a) the state is included in the segment,
(b) a norm of a weight vector of the factorial hidden Markov model corresponding to the state has a large value,
(c) the state is a state where a specific time has elapsed from the start of the segment, and
(d) the state has a large occurrence probability value.
The waveform disaggregation method according to any one of supplementary notes 10-18, comprising an anomaly estimating step of detecting an anomaly of the unit, from the disaggregated signal waveform or a prescribed state.
The waveform disaggregation method according to any one of supplementary notes 33, wherein the anomaly estimating step calculates anomaly level indicating an occurrence degree of anomaly, from the disaggregated signal waveform or the prescribed state, and decides whether or not an anomaly occurs by comparing the anomaly level with a threshold.
The waveform disaggregation method according to any one of supplementary notes 33 or 34, wherein the anomaly estimating step estimates either one or both of a factor in which an anomaly occurs or a state in which an anomaly occurs, based on the signal waveform disaggregated by the estimation section or a prescribed state and compares the anomaly level with a threshold to decide whether or not anomaly occurs.
The waveform disaggregation method according to supplementary note 35, wherein the anomaly estimating step determines priority for a set of the factor and the state, in accordance with an estimated value of a state corresponding to a time at which the anomaly is detected, and
estimates a set of the factor and the state with the priority being high, as either one or both of a factor in which the anomaly occurs and a state in which an anomaly occurs.
The waveform disaggregation method according to supplementary note 36, wherein the anomaly estimating step adopts as criterion for determining the priority, at least one of the followings:
(a) the state is included in the segment,
(b) a norm of a weight vector of the factorial hidden Markov model corresponding to the state has a large value,
(c) the state is a state where a specific time has elapsed from the start of the segment, and
(d) the state has a large occurrence probability value.
The program according to supplementary note 19, causing the computer to execute an anomaly estimating step of detecting an anomaly of the unit, from the disaggregated signal waveform or a prescribed state.
The program according to supplementary note 38, wherein the anomaly estimating processing calculates anomaly level indicating an occurrence degree of anomaly, from the disaggregated signal waveform or the prescribed state, and decides whether or not an anomaly occurs by comparing the anomaly level with a threshold.
The program according to supplementary note 38 or 39, wherein the anomaly estimating processing estimates either one or both of a factor in which an anomaly occurs or a state in which an anomaly occurs, based on the signal waveform disaggregated by the estimation section or a prescribed state and compares the anomaly level with a threshold to decide whether or not anomaly occurs.
The program according to supplementary note 40, wherein the anomaly estimating processing determines priority for a set of the factor and the state, in accordance with an estimated value of a state corresponding to a time at which the anomaly is detected, and
estimates a set of the factor and the state with the priority being high, as either one or both of a factor in which the anomaly occurs and a state in which an anomaly occurs.
The program according to supplementary note 41, wherein the anomaly estimation processing adopts as criterion for determining the priority, at least one of the followings:
(a) the state is included in the segment,
(b) a norm of a weight vector of the factorial hidden Markov model corresponding to the state has a large value,
(c) the state is a state where a specific time has elapsed from the start of the segment, and
(d) the state has a large occurrence probability value.
Number | Date | Country | Kind |
---|---|---|---|
2016-177605 | Sep 2016 | JP | national |
2017-100130 | May 2017 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2017/032704 | 9/11/2017 | WO | 00 |