Industrial and infrastructure systems commonly include supervisory, control, and data acquisition (SCADA) systems. SCADA includes data collection, estimation, and/or decision processes for industrial facilities. SCADA systems are commonly implemented using networks that can include any number of computing devices, sensors, user controls, actuators, etc. SCADA networks can sometimes include internet access, and/or can cover large areas. As a result, SCADA systems can be difficult to secure, and can be significant vulnerabilities for industry and infrastructure installments. There are benefits to improving cybersecurity for such facilities, in particular the networks and devices used to monitor and/or operate infrastructure and industrial systems.
In some aspects, implementations of the present disclosure include a computer-implemented method of training a machine learning model, the method including: generating a random parameter vector; generating a random attack dataset; inputting a plurality of samples of the random attack dataset into a generator; generating, by the generator, a generated attack dataset; training a discriminator using the random attack dataset and the generated attack dataset; and training the generator using the trained discriminators and a loss function.
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein the generated attack dataset includes an attack policy.
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein the attack policy includes a ramp attack.
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein the attack policy includes a sensor attack.
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein the generator includes a deep neural network.
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein the generator is trained with the loss function.
In some aspects, implementations of the present disclosure include a computer-implemented method for training a machine learning model, the method including: receiving a generative model parameter vector; generating sample points from a prior distribution of the generative model parameter vector; generating attack policy parameters simulating simulated attack policy parameters by a system simulation training a generative model by the simulated attack policy parameters.
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein training the generative model includes a deep neural network
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein training the generative model includes selecting a best loss value by.
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein the attack policy parameters include a ramp attack policy.
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein the attack policy parameters include a sine attack policy.
In some aspects, implementations of the present disclosure include a computer-implemented method, wherein the attack policy parameters include a pulse attack policy.
In some aspects, implementations of the present disclosure include a cybersecurity system including: a physical plant a network configured for communications and/or control of the physical plant; a cybersecurity controller operably connected to the network, wherein the cybersecurity controller includes a processor and memory with instructions stored thereon, that, when executed by the processor cause the processor to: receive a trained attack generative model; simulate, by the trained attack generative model, a plurality of attacks on the physical plant; determine an attack effectiveness of each of the plurality of attacks; and control access to the network based on the effectiveness of each of the plurality of attacks.
In some aspects, implementations of the present disclosure include a system, wherein the physical plant includes a networked pipeline system.
In some aspects, implementations of the present disclosure include a system, wherein the networked pipeline system includes a plurality of pressure sensors operably coupled to the network, and controlling access to the network includes securing at least one of the plurality of pressure sensors.
In some aspects, implementations of the present disclosure include a system, wherein the physical plant includes a power grid.
In some aspects, implementations of the present disclosure include a system, wherein the power grid includes a plurality of meters configured to measure the power and frequency of the power grid, and controlling access to the network includes securing at least one of the plurality of meters.
In some aspects, implementations of the present disclosure include a system, wherein the physical plant includes a cyber-physical system.
In some aspects, implementations of the present disclosure include a system, wherein the physical plant includes an industrial facility.
In some aspects, implementations of the present disclosure include a system, wherein the controller further contains instructions that cause the processor to simulate a vulnerability of the physical plant and control the physical plant based on the vulnerability.
In some aspects, implementations of the present disclosure include a system, wherein the controller is further configured to control the network device based on the sampled actions implemented on the system model.
Other systems, methods, features and/or advantages will be or may become apparent to one with skill in the art upon examination of the following drawings and detailed description. It is intended that all such additional systems, methods, features and/or advantages be included within this description and be protected by the accompanying claims.
The components in the drawings are not necessarily to scale relative to each other. Like reference numerals designate corresponding parts throughout the several views.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art. Methods and materials similar or equivalent to those described herein can be used in the practice or testing of the present disclosure. As used in the specification, and in the appended claims, the singular forms “a,” “an,” “the” include plural referents unless the context clearly dictates otherwise. The term “comprising” and variations thereof as used herein is used synonymously with the term “including” and variations thereof and are open, non-limiting terms. The terms “optional” or “optionally” used herein mean that the subsequently described feature, event or circumstance may or may not occur, and that the description includes instances where said feature, event or circumstance occurs and instances where it does not. Ranges may be expressed herein as from “about” one particular value, and/or to “about” another particular value. When such a range is expressed, an aspect includes from the one particular value and/or to the other particular value. Similarly, when values are expressed as approximations, by use of the antecedent “about,” it will be understood that the particular value forms another aspect. It will be further understood that the endpoints of each of the ranges are significant both in relation to the other endpoint, and independently of the other endpoint. While implementations will be described for improving and/or analyzing vulnerabilities in industrial SCADA systems, it will become evident to those skilled in the art that the implementations are not limited thereto, but are applicable for improving the security of communications and control systems in general.
As used herein, the terms “about” or “approximately” when referring to a measurable value such as an amount, a percentage, and the like, is meant to encompass variations of ±20%, ±10%, ±5%, or ±1% from the measurable value.
The term “artificial intelligence” is defined herein to include any technique that enables one or more computing devices or comping systems (i.e., a machine) to mimic human intelligence. Artificial intelligence (AI) includes, but is not limited to, knowledge bases, machine learning, representation learning, and deep learning. The term “machine learning” is defined herein to be a subset of AI that enables a machine to acquire knowledge by extracting patterns from raw data. Machine learning techniques include, but are not limited to, logistic regression, support vector machines (SVMs), decision trees, Naïve Bayes classifiers, and artificial neural networks. The term “representation learning” is defined herein to be a subset of machine learning that enables a machine to automatically discover representations needed for feature detection, prediction, or classification from raw data. Representation learning techniques include, but are not limited to, autoencoders. The term “deep learning” is defined herein to be a subset of machine learning that that enables a machine to automatically discover representations needed for feature detection, prediction, classification, etc. using layers of processing. Deep learning techniques include, but are not limited to, artificial neural network or multilayer perceptron (MLP).
Machine learning models include supervised, semi-supervised, and unsupervised learning models. In a supervised learning model, the model learns a function that maps an input (also known as feature or features) to an output (also known as target or targets) during training with a labeled data set (or dataset). In an unsupervised learning model, the model learns patterns (e.g., structure, distribution, etc.) within an unlabeled data set. In a semi-supervised model, the model learns a function that maps an input (also known as feature or features) to an output (also known as target or target) during training with both labeled and unlabeled data.
Deep learning models, including LLMs, may include artificial neural networks. An artificial neural network (ANN) is a computing system including a plurality of interconnected neurons (e.g., also referred to as “nodes”). This disclosure contemplates that the nodes can be implemented using a computing device (e.g., a processing unit and memory as described herein). The nodes can be arranged in a plurality of layers such as input layer, output layer, and optionally one or more hidden layers. An ANN having hidden layers can be referred to as deep neural network or multilayer perceptron (MLP). Each node is connected to one or more other nodes in the ANN. For example, each layer is made of a plurality of nodes, where each node is connected to all nodes in the previous layer. The nodes in a given layer are not interconnected with one another, i.e., the nodes in a given layer function independently of one another. As used herein, nodes in the input layer receive data from outside of the ANN, nodes in the hidden layer(s) modify the data between the input and output layers, and nodes in the output layer provide the results. Each node is configured to receive an input, implement an activation function (e.g., binary step, linear, sigmoid, tanH, or rectified linear unit (ReLU) function), and provide an output in accordance with the activation function. Additionally, each node is associated with a respective weight. ANNs are trained with a dataset to maximize or minimize an objective function. In some implementations, the objective function is a cost function, which is a measure of the ANN's performance (e.g., error such as L1 or L2 loss) during training, and the training algorithm tunes the node weights and/or bias to minimize the cost function. This disclosure contemplates that any algorithm that finds the maximum or minimum of the objective function can be used for training the ANN. Training algorithms for ANNs include, but are not limited to, backpropagation.
Implementations of the present disclosure include improvements to the security of networked pipeline systems (NPS) and other infrastructure systems. Such systems broadly include “cyber physical systems” that can control the physical elements of a facility or infrastructure. The physical system can be referred to as a “physical plant” and can broadly include machinery, communications systems, networks, and controllers. One example type of infrastructure are networked pipeline systems (NPS) that can play an important role in transporting essential resources such as oil and gas. Increasing integration of information and control techniques in these systems can improve their performance, at the cost of increasing their vulnerability to cyber attacks. Such attacks can have severe consequences on the safety, security, and reliability of pipeline operations. Supervisory control and data acquisition (SCADA) can be used to operate NPS, and provide data collection, state estimation, and decision processes. But SCADA also introduces vulnerabilities. As one example, the state estimation process could be misled maliciously by compromising only a small portion of the IoT-based measurement system [3]. Modifying the control inputs at the automatic control layer can result in catastrophic consequences on physical assets [4].
In order to effectively design and operate security systems, it can be beneficial to model potential attack scenarios and model potential system responses. Such modeling can include generating modeled “attacks,” referred to herein as “attack generation.” Assessing the vulnerability of NPS is a challenging task, as it requires a thorough understanding of the complex interplay between the physical and cyber components of the system and the potential attack scenarios. Existing systems and methods for attack generation and modeling can fail to account for the nonlinear nature of many networked systems (e.g., pipeline systems) and/or require complex or difficult to acquire training data.
Implementations of the present disclosure include improvements to overcome limitations of existing systems for modeling attacks. An example implementation includes an attack generation framework that can integrate physical runtime data of a pipeline system and use a data-driven attack generation approach. The vulnerability analysis can be formulated as determining the presence of feasible attack sets, defined by boundary functions representing the effectiveness and stealthiness of attack signals with respect to the objective and attack detection module. The framework utilizes three data-driven models, including two discriminative models that learn the boundary functions and a generative model that produces elements of the feasible attack set. A loss function defined herein ensures successful attack generation with high probability.
While the implementations described herein are described with reference to industrial and infrastructure systems (e.g., pipelines) it should be understood that implementations of the present disclosure can be used to improve the security and/or operation of any system.
With reference to
For example, the cybersecurity system 100 can include a controller 112 (e.g., a computing device 1100 as described with reference to
It should be understood that the physical plant 150 can be any physical system. For example, the physical plant 150 can be a networked pipeline system including pressure sensors, and the cybersecurity system 100 can be configured to secure the pipeline system by controlling access to the network devices 152, and/or securing parts of the physical systems 154 (e.g., securing one or more of the pressure sensors).
As another non-limiting example, the physical plant can include a power grid, and the physical systems 154 can include one or more meters configured to measure parameters of the power grid (e.g., voltage, frequency, current) at nodes in the power grid. Again, the cybersecurity system 100 can be used to model the vulnerabilities of the power grid, and can control access to the network devices 152 (e.g., meters) and/or physical systems 154 of the power grid to improve the security of the power grid.
With reference to
At step 202 the method includes generating a random parameter vector.
At step 204 the method includes generating a random attack dataset.
At step 206 the method include inputting samples of the random attack dataset into a generator. Optionally the generator can be implemented as a deep neural network.
At step 208 the method includes generating a generated attack dataset using the generator. Optionally, the generated attack dataset can include an attack policy that describes the type(s) of attack that can be used. An example category of attack is an attack on one or more sensors of a physical system (e.g., by injecting false sensor data into a network). Non-limiting examples of this type of attack include ramp attacks, where false data is ramped up or down over time.
At step 210 the method includes training a discriminator using the random attack dataset and the generated attack dataset.
At step 212 the method includes training the generator using the trained discriminator and a loss function. An example loss function that can be used is:
Additional description of loss functions is provided with reference to Example 1, hereto.
At step 222 the method includes generating sample points from a prior distribution of the generative model parameter vector.
At step 224 the method includes generating attack policy parameters. Non-limiting examples of attack policy parameters include a ramp attack policy, sine attack policy, and pulse attack policy. The present disclosure contemplates that combinations of attack policy parameters can also be used.
At step 226 the method includes simulating simulated attack policy parameters by a system simulation.
At step 228 the method includes training a generative model by the simulated attack policy parameters. Optionally the generative model is a deep neural network, as described in example 2 herein. The deep neural network can optionally select a best loss value by the function: Jbest(i+1)=min{Jbest(i),J(θg(i+1))}.
It should be appreciated that the logical operations described herein with respect to the various figures may be implemented (1) as a sequence of computer implemented acts or program modules (i.e., software) running on a computing device (e.g., the computing device described in
Referring to
In its most basic configuration, computing device 1100 typically includes at least one processing unit 1106 and system memory 1104. Depending on the exact configuration and type of computing device, system memory 1104 may be volatile (such as random access memory (RAM)), non-volatile (such as read-only memory (ROM), flash memory, etc.), or some combination of the two. This most basic configuration is illustrated in
Computing device 1100 may have additional features/functionality. For example, computing device 1100 may include additional storage such as removable storage 1108 and non-removable storage 1110 including, but not limited to, magnetic or optical disks or tapes. Computing device 1100 may also contain network connection(s) 1116 that allow the device to communicate with other devices. Computing device 1100 may also have input device(s) 1114 such as a keyboard, mouse, touch screen, etc. Output device(s) 1112 such as a display, speakers, printer, etc. may also be included. The additional devices may be connected to the bus in order to facilitate communication of data among the components of the computing device 1100. All these devices are well known in the art and need not be discussed at length here.
The processing unit 1106 may be configured to execute program code encoded in tangible, computer-readable media. Tangible, computer-readable media refers to any media that is capable of providing data that causes the computing device 1100 (i.e., a machine) to operate in a particular fashion. Various computer-readable media may be utilized to provide instructions to the processing unit 1106 for execution. Example tangible, computer-readable media may include, but is not limited to, volatile media, non-volatile media, removable media and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. System memory 1104, removable storage 1108, and non-removable storage 1110 are all examples of tangible, computer storage media. Example tangible, computer-readable recording media include, but are not limited to, an integrated circuit (e.g., field-programmable gate array or application-specific IC), a hard disk, an optical disk, a magneto-optical disk, a floppy disk, a magnetic tape, a holographic storage medium, a solid-state device, RAM, ROM, electrically erasable program read-only memory (EEPROM), flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices.
In an example implementation, the processing unit 1106 may execute program code stored in the system memory 1104. For example, the bus may carry data to the system memory 1104, from which the processing unit 1106 receives and executes instructions. The data received by the system memory 1104 may optionally be stored on the removable storage 1108 or the non-removable storage 1110 before or after execution by the processing unit 1106.
It should be understood that the various techniques described herein may be implemented in connection with hardware or software or, where appropriate, with a combination thereof. Thus, the methods and apparatuses of the presently disclosed subject matter, or certain aspects or portions thereof, may take the form of program code (i.e., instructions) embodied in tangible media, such as floppy diskettes, CD-ROMs, hard drives, or any other machine-readable storage medium wherein, when the program code is loaded into and executed by a machine, such as a computing device, the machine becomes an apparatus for practicing the presently disclosed subject matter. In the case of program code execution on programmable computers, the computing device generally includes a processor, a storage medium readable by the processor (including volatile and non-volatile memory and/or storage elements), at least one input device, and at least one output device. One or more programs may implement or utilize the processes described in connection with the presently disclosed subject matter, e.g., through the use of an application programming interface (API), reusable controls, or the like. Such programs may be implemented in a high level procedural or object-oriented programming language to communicate with a computer system. However, the program(s) can be implemented in assembly or machine language, if desired. In any case, the language may be a compiled or interpreted language and it may be combined with hardware implementations.
The following examples are put forth so as to provide those of ordinary skill in the art with a complete disclosure and description of how the compounds, compositions, articles, devices and/or methods claimed herein are made and evaluated, and are intended to be purely exemplary and are not intended to limit the disclosure. Efforts have been made to ensure accuracy with respect to numbers (e.g., amounts, temperature, etc.), but some errors and deviations should be accounted for. Unless indicated otherwise, parts are parts by weight, temperature is in ° C. or is at ambient temperature, and pressure is at or near atmospheric.
An example implementation of the present disclosure includes an attack generation framework for evaluating the vulnerability of nonlinear networked pipeline systems.
As used in the present example, n,
+ denotes the space of real vectors of length n and positive real numbers respectively. The probability triple is denoted by (Ω,
, PZ) [9], where Ω is a sample space containing the set of all possible outcomes,
is an event space, and Pz is the associated probability functions of the events in
. The study used the symbol
to denote the expected value operator. An undirected graph
(V, ε) contains vertices, denoted by ε={v1, v2, . . . , vn}, and edges, denoted by ε∈V×V.(v1, v2) and (v1, v2)∈ε represents an edge in
. Consequently, the adjacency matrix, denoted by A(
), is a square matrix of size |V|, defined as [10]
An example implementation was studied with a model of a networked pipeline system, including the physical plants, network dynamics, a supervisory control and data acquisition (SCADA) system, and an attack model.
A nonlinear creep flow model was used to describe the gas flow in a pipeline segment [11]:
and wi is the mass flow at pipeline i, Δxij denotes the length of pipeline between node i and node j, denote the set of neighborhood pipelines of the pipeline i. Let
ωij is the compressor rotor angular velocity [rad/s], qij is the mass power flow [kg/s], ϕ(ω, q) is the compressor characterization, uij is the mechanical torque [N-m] applied to the rotor shaft or inertia J0
A supervised control framework was employed to stabilize pipeline pressure and regulate mass flow, compromising a supervised control layer and an automatic control layer. The supervised control layer contains an estimator, a bad data detector (BDD), and a supervised controller. The pipeline system is equipped with hundreds of smart pressure and flow meters, the goal of the estimator at the supervised control layer is to fuse the meter readings and estimate the states of interest p. The measurement model is given as
Then an unscented Kalman Filter (UKF) is utilized to perform sensor fusion and state estimation. Firstly, following standard unscented transformation, the study used 2n+1 sigma points to approximate the state p with assumed mean
The corresponding weights for the sigma points are given as W0m=λ/(n+λ), W0c=W0m+(1−α2+β), Wi=½(L+λ), and λ=α2(n+κ)−n represents how far the sigma points are away from the state, κ≥0, α∈(0,1], and β=2 is the optimal choice for Gaussian distribution. Assuming pk−1˜(
Linearizing the models around the equilibrium point (peq, weq) and discretizing with a fixed time step Ts yields:
Then, a model predictive controller of horizon length h is utilized to generate the reference mass flow wr:
Thus, the reference mass flow is given as wkr=Δw*[k]+weq. Next, at the automatic control layer, a PID controller is used to control the compressors to track the reference mass flow.
In the example study, attacks can be injected through the sensing process and actuation process, and it can be modeled as [15], [14]
To characterize the effect of attacks in the system (11), effectiveness and stealthiness are two common criteria [3]. The study used l1: m×
p→
, l2:
m×
p→
to denote functions evaluating the effectiveness and stealthiness of the attacks respectively. Then a feasible set of attacks S is defined with given thresholds of effectiveness and stealthiness τE, τS:
Most research uses the estimation error to represent the effectiveness of sensor attacks such that l1=∥{circumflex over (x)}i−xi∥[13], [14]. The stealthiness function often employs the BDD function l2=D, where D is defined in (7).
Consequently, the attack generation problem is finding a generative model of the form
The study introduced a solution to search for feasible attacks in the set defined in (12) using only the runtime data of NPS. The data-driven attack generation framework is illustrated in
The generator G(z, θ) can be a deep neural network learning the distribution of feasible attacks from sampling noise. It can optionally be trained with the loss function
Successful training of the generator can require knowledge of l1 and l2 functions. However, it is hard to obtain their closed-form expression. Data-driven approximation offers a practical solution, as it only requires runtime data instead of a high-fidelity model and provides easy calculation of gradients. The study used two deep regression neural networks to approximate l1 and l2:
They are both trained with the mean-square-error (MSE) loss function given N runtime effectiveness metric data l1(e) with the corresponding injection of attacks e:
The generator maps random samples z˜Pz to the parameter vector I∈p of a pre-defined attack policy π(I). Given I∈
p, the attack policy π(I) is a deterministic time sequence of the injected attacks given by
It can include the start time of the attack injection t0(I), the duration of the attack injection T(I), and the attack profile g(I): [t0(I)t0(I)+T(I)]→n. The specific attack policy is assumed to be predetermined for the purpose of the example implementation studied.
Biased learning can be a concern in deep generative models [16], particularly when trained with multiple interrelated data-driven models. The discriminators might learn biased l1 and l2 functions based on a limited space of attacks covered by the generative model. This in turn affects the training of the generator network. To improve and address biased learning issues, implementations of the present disclosure can incorporate random and generated attack datasets into the discriminator training process. An example training procedure can include: Receiving Hyperparameters: τE, τS, α, π, s;
A simulation was performed to evaluate the proposed attack generative model on a 4-node pipeline system, as depicted in
The topology of the network can be represented by the adjacent matrix:
The pipeline model parameters were chosen as c=330 m/s, D12=0.8 m, D23=0.5 m, D34=1.5 m, Δx12=Δx23=Δx34=10 m, f12=f23=f34=0.0025. The compressor model parameters are chosen as J0=47.7465, ρr22=0.0951, Tin cp=13322.3, kf=0.05, r1=20, Acr1ρin=0.5834, γ=1.2, and Ac/rc=0.0146
In the simulation, the study targeted the equilibrium point peq=[199 194 54 50]T, corresponding to the equilibrium point of the mass flow weq=[−20 −5 −5 13]T.
The objective of the attack generation task is to fool UKF into giving biased estimates {circumflex over (x)}a leading to small r but big e. The feasible attack set is given particularly as:
Here ē is the maximum magnitude of the attack signal. The attack injection start time t0∈[tl tr] and the duration time T∈[Tl Tr]. The study use e=0.5, tl=20 s, tr=40 s, Tl=0 s, Tr=100 s and the simulation time for generator training is set as 200 s. These exact values of t0, T, e at runtime are specified by the parameters I∈3 generated by the generator.
Next, the study implemented the example algorithm with τE=10, τS=0.2, α=0.8, s=0.01 and π given by 25. The generator employed was a deep neural network composed of 4 layers, ReLU, ReLU, Tanh, Sigmoid, with 500, 1000, 500, 3n neurons respectively, where n is the number of nodes under attack. The input size was set as 10. The stealth net includes of 3 ReLU hidden layers and Sigmoid output layer, and the effect net includes of 3 ReLU hidden layers and Linear output layer. The numbers of neurons at layers are 500, 1000, 500, 1 for both effect net and stealth net. They are trained by (20) and 19) respectively. Adams optimizer, in Matlab deep learning toolbox, is used with the learning rate 0.0002, gradient decay factor 0.5, and square gradient decay factor 0.999. The study trained the framework in 5 epochs. In each epoch, the generator trained 10 batches of 5000 batch size, while the discriminators are trained 200 batches of 1000 batch size.
After each epoch of training, the performance of the attack generator is tested in a real system simulation, as shown in
Finally, the study demonstrated the effectiveness of the proposed attack generation method through a time-series performance of generated feasible attacks in
The example described herein implements a data-driven attack generative model that can be used for NPS vulnerability analysis. The framework includes three interactive data-driven models: two discriminative models to evaluate attack signals and a generative model to generate feasible attacks. The loss function disclosed ensures successful attack generation.
It should be understood that the example in the study is non-limiting, and that other loss functions and/or applications of implementations of the present disclosure are contemplated. For example, the present disclosure contemplates interactive training between the proposed attack generator and a supervised learning-based attack detector, which can lead to a complete automatic exploration of the vulnerability space and the development of a perfect attack detector for the system.
A second study was performed on a second example implementation of the present disclosure. The second study included improvements to modeling and design of cyber-physical systems to reduce vulnerability to cyber attacks. As used in the present study, the summation symbol used here is a generalized summation for the appropriate random variable. If the random variable is continuous, then the summation symbol will be considered an integral sign. If the random variable is discrete, then the summation symbol will be the traditional summation of countable quantities. Let x1, x2∈. Then, ReLU(x1)≤ReLU(x2) implies that x1≤x2. Otherwise, x1≤0 and x2≤0—in which case ReLU(x1)=ReLU(x2)=0.
Modern cyber-physical infrastructure systems (CPIS) are built based on the seamless integration of physical devices and computer systems, facilitating data exchange and control of physical processes [1A]. As the complexity and scale of CPIS increase, network structures are utilized to leverage the resources and capabilities of subsystems, resulting in improved functionality and performance [2A]. In such NCPS, the physical agents, sensors, and actuators are tightly interconnected or coupled through communication networks to gather and analyze data and perform physical actions in response to high-level commands or unwanted external influence. NCPSs are used in various applications, such as smart power grids [3A], industrial pipeline systems [4A], and transportation systems [5A]. The networked and distributed integration of computer systems and physical devices provides numerous benefits, including improved efficiency, reduced costs, and increased safety [6A]. However, networked information transmission presents significant security and privacy challenges as they are vulnerable to cyber-attacks and data breaches [7A]. Therefore, the development of secure and reliable NCPS is essential to ensure their safe and effective operation [8A].
Assessing the vulnerability of NCPS is a challenging task that requires a deep understanding of the complex interplay between the physical and cyber components of the system and the potential attack scenarios. Various techniques to address the vulnerability analysis problem in cyber-physical systems (CPS) exist, but are limited. Such techniques can include mathematical and analytical methods, as well as data-driven and simulation-based approaches. Early researchers incorporated the full system model into the maximization program to generate a feasible attack [9A]-[11A]. In [9A], a false data injection attack (FDIA) was studied against the least-square estimator with a residual-based bad data detector (BDD). The feasibility of FDIA against the Kalman filter with χ{circumflex over ( )}2 detector was studied in [10A]. One study defined vulnerability under sensor attacks as the boundedness of the estimation error and derived sufficient and necessary conditions through analysis of the system's reachable zero dynamics [12A]. A sufficient and necessary condition for insecure estimation under FDIA was also derived for a networked control system in [11A]. To develop more pragmatic attack generation strategies, several constraints are incorporated to capture the attacker's limitations such as limited access to sensors [9A], shallow knowledge of system dynamics [13A], and incomplete knowledge of implemented state estimators [14A]. However, these model-based approaches primarily target linear CPS.
Given the nonlinear nature of NCPSs and the challenges associated with obtaining high-fidelity models, the corresponding vulnerability analysis problem is significantly more complex [15A]. One approach is to analyze vulnerability based on domain knowledge and extensive simulations. For instance, [16A] combined a stochastic adversarial model with a simulation model of interdependent gas-electricity delivery infrastructures to explore operational disruptions caused by cyber-attacks. Another study in [17A] employed hierarchical Limited Stochastic Petri nets and power system network topology to simulate intrusion attack scenarios. Furthermore, a dynamic network security vulnerability assessment method was developed for SCADA systems, taking into account software vulnerabilities and deployed network security defense measures [18A]. However, simulation-based approaches may only be possible for a specific system with specific attack.
Alternatively, data-driven approaches can provide more general solutions. In [19A], [20A], generative adversarial networks (GANs) were trained to learn from existing feasible attack datasets, but their performance relies heavily on the availability of non-generative attack datasets and the representative quality of the training data. Another approach in [15A] utilized data-driven models to approximate the system model, thereby reducing the complexity of solving the optimization-based attack generation problem. However, the optimization-based generator produces a specific attack rather than exploring the vulnerability space.
While existing approaches have been shown to be effective in various applications, they have limitations in capturing the nonlinear nature of NCPSs, relying on exact system model information or prior attack datasets, and being unable to explore vulnerability space. The study disclosed an example implementation including an attack generative framework that overcomes these limitations by integrating knowledge of the underlying physics of NCPS and a data-driven attack generative methodology. A tailored loss function for training the proposed model is given, with backing theoretical guarantees for convergence and probabilistic coverage of the vulnerability space. The example implementation can be applicable to both linear and nonlinear systems without relying on exact model knowledge assumption or prior attack datasets. The example implementation can include a unified attack generation framework for any NCPS. The proposed attack generation system was validated on an IEEE 14-bus system and various gas pipeline systems.
As used in the present example, n,
+ and
2 denote the Euclidean space of dimension n, set of all positive real numbers, and the Hilbert space of all square summable signals of dimension n respectively. Normal-face lower-case letters (e.g. x∈
) are used to represent real scalars, bold-face lower-case letters (e.g. x∈
n) represent vectors, while normal-face upper-case letters (e.g. X∈
m×n) represent matrices. The study used xi to denote the i th element of the vector x∈
n and x(i) the vector signal x∈
n at time i. dist(θ, Θ) is used to denote the Euclidean distance of a vector θ∈
n to a set Θ*g⊂
n.
The probability triple is denoted by (Ω, , Pz), where Ω is a sample space containing the set of all possible outcomes,
is an event space, and Pz is the associated probability functions of the event z in
. The study uses the symbol
to denote the expected value operator. The ReLU function is given by
ReLU(x)=max(0,x)
(V, ε) denotes an undirected graph with vertices, denoted by ε={v1, v2, . . . , vn}, and edges, denoted by ε∈V×V.(v1, v2) and (v1, v2)∈ε represents an edge in
. Consequently, the adjacency matrix, denoted by A(
), is a square matrix of size |V|, defined as [21A]
The support of a vector x∈n is defined as supp(x)
{i⊆{1, . . . , n}|xi≠0}. Σk
{x∈
n∥supp(x)|≤k} denotes the set of k-sparse vectors. The study uses ⊙ to denote element-wise multiplication, i.e.
The study uses ⊗ to denote Kronecker product, i.e. for B∈m×n,
The study considers a networked CPS under attack. The networked CPS contains n nodes/agents with communication and computation capability.
The example implementation included a system dynamical model. The physical process is modeled by the network dynamics given by:
Let x=[x1T x2T . . . xnT]T∈m
Since the underlying graph is undirected, it is shown that the nodal interaction matrix Ψ satisfies the skewsymmetric properties; Ψ(x)T=−Ψ(x) and ηTΨ(x)η=0 for all X∈m
n. Consequently, the entire network dynamics are given as
Let a denote the index set of actuation nodes, then the actuation state vector zi=0 if i∉
a, otherwise it follows an actuation dynamics described by the control-affine dynamical model, subject to actuation attacks:
Next, a measurement model of equipped IoT sensors under attack is given by
The example implementation further included a nominal control design. Here, the study provides a control design for the nominal system without attacks (i.e. eu=0, ey=0). The materials in this subsection can be inferred from standard literature on nonlinear control design and are presented for the sake of completeness. The control design comprises 2 layers; (1) the low-level actuator control (Level 1), and (2) the supervisory reference generator (Level 2).
The low-level controllers are closed-loop schemes ui: m
p designed to ensure that the actuator state zi tracks a given reference zir∈
p such that
Thus, there exists a stabilizing controller of the form u=K(V({tilde over (z)}))—for example using the Sontag formula [26A]
are the Lie derivatives of V along f and g respectively.
The supervisory control generates the state reference zr for the low-level controllers in order to regulate the network states x at the equilibrium point. Consistent with the time-scale separation inherent in a typical NCPS, the study assumes that the lowlevel control layer is much faster than the supervisory layer. Thus, for the purpose of designing a supervisory target generator, the study linearized the system dynamics about the operating point [27A]. The operating point is a known special equilibrium point for which the nominal operation of the CPS is designed. Examples of operating points include the nominal frequency and bus voltage of an AC transmission system (frequency of 60 Hz and voltage ranging from 230 KV to 500 KV for a typical transmission system in the United States). Although the nonlinear dynamics in (3) would have multiple equilibrium points, the study uses the known operating point as the equilibrium point of interest. Let xeq be the equilibrium point of interest corresponding to the equilibrium points of inputs ueq. Next, a linear model is used to approximate the network dynamical model in (3) around xeq:
Next, given a sample time step Ts, a discrete form of the linear model in (9) is given as:
Consequently, a model predictive control (MPC) scheme is utilized to regulate the network states at the equilibrium point xeq:
The goal of the MPC program is to minimize the deviation from the operating point and the corresponding control energy using weight matrices M1, M2 respectively. The discrete error dynamics is imposed as an equality constraint. The initial constraint {tilde over (x)}(0)={tilde over (x)}t and terminal constraint {tilde over (x)}(k)=0 are imposed as another equality constraints. Since the network state x is not measurable, the regulation error is given by {tilde over (x)}(t)={circumflex over (x)}(t)−xeq, where the estimate of the network states {circumflex over (x)}t is obtained using the following state estimator, given sensor measurements y.
The state estimator is a minimizer of the difference between model measurements and sensing readings in T horizon, according to the program
As the estimator estimates the network states from the measurements, a BDD monitors the state estimation process to detect any false inputs. It is defined as a function D∈: m
m→
+ mapping from the state estimates to a detection residual (i.e. 1-norm or 2-norm residual-based detectors [9A], [28A]) or detection likelihood (i.e. χ2 detector [10A], [15A]).
The example implementation further included analyses of a vulnerability analysis problem.
Since obtaining a high-fidelity model for complex networked systems is challenging even for experienced practicing engineers and operators, the system models described in the present example are assumed to be unknown. However, for the purpose of vulnerability analysis, the study assumed that all measurement and actuation channels can be accessed. In other words, the study assumes all actuators and sensors are potential targets for attacks in order to cover all possible vulnerable scenarios. Additionally, the study assumes that the effectiveness and stealthiness of the injected attacks could be measured. Stealthiness refers to the potential to bypass BDD, while effectiveness represents the closeness to the intended degradation of system performance [11A]. Various metrics have can be used to measure effectiveness and stealthiness. For example, estimation error was used as the effectiveness metric while the estimation residual was used as the stealthiness metric in [10A], [14A]. [28A] used a critical measurement, such as the sum of the water levels in a water tank system, as the effectiveness metric. The present example can use any of these metrics, and/or allow the system operator to select metrics and consider them as hyperparameters to the vulnerability analysis problem. Next, the study defines a vulnerable NCPS as follows.
Definition 1 (Vulnerable System). For the NCPS in (3), (5) and (6), let l1: 2→
, l2:
2→
be the corresponding Lipschitz effectiveness and stealthiness functionals respectively. Consider a vulnerability set
where τE, τS are the corresponding effectiveness and stealthiness thresholds. Then
Specifying the l1 and l2 as Lipschitz functionals is justifiable as they are often given by energy-like functions expressed in terms of certain norms and/or inner products. This can be achieved by either employing a linear model approximation [29A] or utilizing data-driven universal approximation [15A].
To assess the vulnerability of the NCPS in (3), (5) and (6), which is equivalent to certifying the nonemptyness of , it is sufficient to generate any feasible eu, ey. Therefore, the vulnerability analysis can be formulated as an attack generation problem [30A]. The study shows that the domain of
(τE, τS) is a function space, thereby rendering the attack generation problem infinite dimensional. To circumvent this, the study employs an attack policy, which is a parameterized basislike signal vector used as a heuristic to simplify the search space for the generative model. This essentially transforms the original infinite-dimensional search space into a finitedimensional parameter search. Next, the study formalizes this definition of attack policy.
Definition 2 (Attack Policy). An attack policy π(ϕ) is a mapping from policy parameter vector ϕ∈n
Thus, this results in a more conservative vulnerability set
Definition 3 (Generative model). A generative model is a mapping of the form
Theorem 1. Given a set whose boundary is of measure zero. If
Proof. By definition of the expectation operator, for any random variable x,1
Since D(x)≥0, it follows that
In addition, since D(x)>1 for all x∉,
[D(x)] can be further lower bounded as
Replacing x with G(z, θ) yields
[D(G(z,θ))]≥Pr{G(z,θ)∉
}
Using the inequality in (17), it follows that
Consequently, for a given threshold η∈(0,1), the study defined the feasible parameter set Θg(η):
By utilizing any function D which satisfies the discrimination property in (18), the nonconvex vulnerability analysis problem can be transformed to a training problem for the generative model G(z, θ). For this, the study considers the example loss function
Theorem 2. Given η∈(0,1), if Θ*g(η) is nonempty, then
Proof. Let
Then, given any arbitrary θ1∈n
Now, due to the order preservation of the ReLU function, either
(21) holds for all
it follows that
The training convergence of the generative model with the loss function in (20), using the specific discrimination function of the form:
D(ϕ)=
First, the study analyzed the convergence guarantee when l1, l2 are known exactly. However, calculating the gradient of such a nonlinear networked dynamical system can be computationally expensive. As a result, the study used universal approximators to learn the functions l1 and l2 from runtime data. The study also shows that the convergence property is preserved using the approximated versions {circumflex over (l)}1 and {circumflex over (l)}2. To obtain a parameter vector within the feasible set Θ*g(η), the study employed the iterative scheme:
Proposition 1. Suppose there exists a compact subsets Θg⊂n
n
n
and l2∘π:
n
satisfy the Lipschitz conditions:
is Lipschitz continuous. Thus, for any compact set Θg⊂n
Thus, for all θ1, θ2∈Θg, the following follows from the Lipschitz continuity of the ReLU function
Remark 2. As an example of (22), it is not hard to check that
Next, the study presents results on the asymptotic convergence of the scheme in (23).
Theorem 3. Consider a generative model G(z, θg) and discriminator D(ϕ) which satisfy the Lipschitz conditions in Proposition 1. Given any initial guess θg(0)∈Θg. Suppose that the feasible parameter set satisfies the inclusion Θ*g(η)≤Θg, and that the step size sequence of the iterative scheme in (23) is chosen to satisfy the nonsummable square-summable property
Iterating the last inequality over {0, . . . , i+1} yields
Since ∥θg(i+1)−θ*g∥2≥0, it follows that
On the other hand, the following inequalities hold
From Proposition 1, it is seen that the loss function J(θ) is uniformly Lipschitz over Θ*g(η) with the constant LD(L1+L2)LG. Thus, the subgradient h(i) can be upper bounded as:
leading to the inequality:
Remark 3. Since ϵ is arbitrary and J(i)≥J(θ*), it follows that limi→∞Jbest(i)=J(θ*), which according to the feasibility set defined in (19), implies that limi→∞θg(i)∈Θg(η) for any given η.
Next, to reduce the computational burden of the gradient evaluation, the study approximated the functions l1∘π, l2∘π using universal approximators and develop a corresponding training algorithm for the generative model. The two data-driven models {circumflex over (l)}1, {circumflex over (l)}2 are assumed to satisfy the universal approximation properties [31A], [32A]: For any ϵ1, ϵ2∈+, there exists parameter vectors β1 and β2 such that
Corollary 3.1 (Convergence with Approximated {circumflex over (l)}1, {circumflex over (l)}2). Suppose the data-driven models {circumflex over (l)}1(θE(i)), {circumflex over (l)}2(θS(i)) satisfy the universal approximation properties in (37), where θE(i) and θS(i) are optimal parameters at epoch i. Then, by following Algorithm 1 with the nonsummable square-summable property in (31), for any ϵ>0,
and Nϵ is the constant given in (33).
Proof. Using the data-driven models {circumflex over (l)}1 and {circumflex over (l)}2, the iterative scheme becomes
Essentially, the above theorem shows that, with the uncertainties on the l1, l2 functions, the feasible set Θ*g(η) can still be reached but with a particular amount of increased convergence time. The increase in the convergence time depends on the choice of step size and the uncertainty bounds.
The example implementation included case studies for a power grid system and gas pipeline network. The case studies included attack policies that can be used in implementations of the present disclosure.
The study considered a structure of attack policy as:
It includes the start time of the attack injection t0∈[tl, tu], the duration of the attack injection T∈[Tl, Tu], and the attack profile e(θa, t). [tl, tu] and [Tl, Tu] are the pre-defined interval of the start time and duration of the attack injection respectively. Examples of such attack policies can be found in
The attack injection time t0 and the duration time T are determined as
The attack profile functions are determined by two parameters e1∈[e1, ē1], e2∈[e2, ē2], where [e1, ē1] and [e2, ē2] are also
Then, for RA policy, the attack profile function is given as:
For SA policy, it is given as
For PA policy, it is given as
The study considered a power grid system comprising ng generator buses and nl load buses. An example on IEEE 14-bus system (ng=5, nl=14) is used, as shown in
The study included an example system's dynamical models and the nominal control setup. Without loss of generality, several assumptions are admitted [34A]:
and Pgnω(θ), Plnω(θ) are the active power vector at generator buses and load buses, respectively. Furthermore, the nonlinear swing dynamics of generators is described by [34A]:
Since the network power dynamics has already been at its steady state, the supervisory controller in (11) is not required, and the reference active power for generators can be determined as
By ordering the buses such that the generator nodes appear first, and linearizing the model in (52), the study obtained:
in which θl is the bus angle of load buses. Combining (51) and (53) yields
Then substituting (55) into the generator dynamical equation (50) yields
where L is the observer gain satisfying that Ad−LC has all the eigenvalues inside the unit circle.
Then the PI controller:
Next, the study discussed the implementation of the proposed attack generator on this IEEE 14-bus system. The attacks are injected through the measurements. The total simulation time is set as 8 s, thereby Tsim=8/Ts=800 time steps in total. The study choose the ratio of the mean control error of the generator rotor angles to the maximum nominal3 generator angles to be the effectiveness index
The three attack policies, shown in
All three attack policies are used separately in the training process of Algorithm 1 shown in
The same stealthiness and effectiveness thresholds are chosen: ϵ=0.05 and α=49%, so the feasible attack set is defined as
For Algorithm 1, other hyperparameters are chosen as η=0.8 and s=0. (0.05, 49%) to the total testing samples for the epochs. It shows that the generator with SA policy converges faster than the other two, and the generator with PA policy is the slowest, likely due to its complexity.
Next, the study evaluated the performance of the generated attacks with different attack policies by injecting them into the system's time-series simulation. As shown in
The study further considered gas pipeline systems equipped with IoT sensors and actuators, a supervisory controller, and automatic controllers, as shown in
A nonlinear creep flow model is used to describe the gas flow in a pipeline segment [35A]:
In addition, a dynamical model of the compressor between two nodes i and j is given by [36A]:
ωij is the compressor rotor angular velocity [rad/s], qij is the mass power flow [kg/s], ϕ(ω, q) is the compressor characterization, uij is the mechanical torque [N-m] applied to the rotor shaft or inertia J0
The supervisory target generator is given in (11) to regulate the node pressures around its equilibrium point at a lower frequency. The study assumed all node pressures and mass flows could be directly measured subject to noise v:
composes of the node pressure states, the wellhead supplied mass flow and the demand mass flow. An unscented Kalman filter (UKF) [37A] is used to estimate the states from the noisy measurements. The study tested three different pipeline systems with distinct network topologies, namely the Linear topology, Tree topology, and Cyclic topology, as illustrated in
Next, the study discusses the implementation of the proposed attack generation system in these pipeline systems. The attacks are injected through the measurements, then the measurement model in (71) becomes
The total simulation time is set as 200 s, thereby Tsim=200/Ts=20000 time steps in total. To train the attack generator, the example implementation used the ratio of the mean value of the node pressure control error to a constant nominal pressure as the effectiveness metric:
This case study focused on testing the Ramp attack policy, as defined in (45), with the pre-defined range of injection start time t0∈[0.1 Tsim, 0.2 Tsim], injection duration time T∈[1e-5, 0.5 Tsim] and two attack value function parameters e1∈[0.1 Tsim, Tsim] and e2∈[0,0.5*y]. The input of the attack value function (45) is given in (43), where the input of the attack policy g is the output of the generator. For linear topological pipeline system, it has 12 measurements, so the output size of the associated generator is 4*12=48. Similarly, the output size of the generator for the tree topological pipeline system is 4*14=56 and the output size of the generator for the cyclic topological pipeline system is 4*16=64. The input sizes of the generators are the same and equal to 10, and they all have 4 layers: ReLU (500 neurons), Re LU (1000 neurons), Tanh (500 neurons), Sigmoid. The stealth net includes of 3 ReLU hidden layers and a Sigmoid output layer, and the effect net includes of 3 ReLU hidden layers and a Linear output layer. The numbers of neurons at hidden layers are 500, 1000, 500 for both the effect net and stealth net. In each training epoch, the discriminators are trained in 20 batches with a size of 1000, and the generator is trained in 5 batches with a size of 5000. The same stealthiness and effectiveness thresholds are chosen for all three topological pipeline systems: ϵ=0.02 and α=65, so the feasible attack set is defined as:
For Algorithm 1, other hyperparameters are chosen as η=0.8 and s=0. All generators converge in the first 2 epochs. Therefore, in order to show the convergence process of the generative system, the study plotted the testing results after each batch of training in the first 2 epochs.
Next, the study evaluated the performance of generated feasible attacks by injecting them into the time-series simulation of those three pipeline systems. The results are presented in
The study developed a data-driven attack generation system accessing the vulnerability of nonlinear NCPS. It does not require the model knowledge of the system but only runtime data. The framework comprises three interactive models: two discriminative models that assess attack signals and a generative model that generates feasible attacks. To ensure a high probability of successful attack generation, the study disclosed a loss function and a training algorithm with convergence guarantee. The implementations described herein are highly versatile and can be applied to various NCPS. It can enable the learning of stealthiness boundaries while exploring effectiveness, which holds immense potential for use in resilient co-design with learning-based BDDs.
The study further contemplates that implementations of the present disclosure can automatically find an optimal attack policy using an algorithm similar to policy learning and/or include an automatic approach to identify the thresholds of effectiveness and stealthiness, ensuring that the feasible attack set is not void. Implementations of the present disclosure can include additional features not described as part of algorithm 1 shown in
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims.
The following patents, applications, and publications, as listed below and throughout this document, describes various application and systems that could be used in combination the exemplary system and are hereby incorporated by reference in their entirety herein
This application claims the benefit of U.S. provisional patent application No. 63/609,587 filed on Dec. 13, 2023, and titled “SYSTEMS AND METHODS FOR ASSESSING THE VULNERABILITY OF CYBER-PHYSICAL SYSTEMS,” the disclosure of which is expressly incorporated herein by reference in its entirety.
This invention was made with government support under Grant number DECR0000005 awarded by the Department of Energy. The government has certain rights in this invention.
Number | Date | Country | |
---|---|---|---|
63609587 | Dec 2023 | US |