REINFORCEMENT LEARNING ALGORITHM-BASED PREDICTIVE CONTROL METHOD FOR LATERAL AND LONGITUDINAL COUPLED VEHICLE FORMATION

Information

  • Patent Application
  • 20240083428
  • Publication Number
    20240083428
  • Date Filed
    July 13, 2023
    9 months ago
  • Date Published
    March 14, 2024
    a month ago
Abstract
A reinforcement learning algorithm-based predictive control method for lateral and longitudinal coupled vehicle formation includes S1, combining a 3-DOF vehicle dynamics model that takes into account a nonlinear magic formula tire model with a lane keeping model and establishing a vehicle formation model; S2, constructing a distributed control framework and designing a local predictive controller for each following vehicle based on the vehicle formation model under the control framework; S3, using a reinforcement learning algorithm to solve the optimal control strategy of the local predictive controller, and applying the optimal control strategy to the target following vehicle. The present application completes the lateral and longitudinal coupled modeling of vehicle formation and considers the nonlinear characteristics of tires. In addition, the present application also transforms the global optimization problem of vehicle formation into a local optimization problem of each following vehicle.
Description
CROSS REFERENCE TO THE RELATED APPLICATIONS

This application is based upon and claims priority to Chinese Patent Application No. 202211087600.7, filed on Sep. 7, 2022, the entire contents of which are incorporated herein by reference.


TECHNICAL FIELD

The present application belongs to the field of automobile control technology, specifically to a reinforcement learning algorithm-based predictive control method for lateral and longitudinal coupled vehicle formation.


BACKGROUND

In the past few decades, the exponential growth in the number of vehicles has brought great challenges to energy security and traffic safety. According to the National Highway Traffic Safety Administration, about 84% of traffic accidents are attributed to human errors. Vehicle formation can significantly reduce traffic accidents caused by driver fatigue and mishandling, thereby improving road safety. In addition, vehicle formation (especially heavy truck formation) can reduce air drag between vehicles, resulting in lower emissions, fuel consumption, and increased road capacity. These potential benefits have led to an increased interest among scholars in vehicle formation control.


The research on vehicle formation mainly includes longitudinal control and lateral control, which involve:


The goal of longitudinal control is to align the desired speed of the vehicle and to maintain the desired spacing between neighboring vehicles.


The task of lateral control is to steer the vehicle within the designated lane.


The existing research on vehicle formation control mostly adopts the decoupling control method, where separate lateral and longitudinal controllers are designed to realize lateral lane keeping and longitudinal speed tracking respectively. Specifically, the lateral control methods mainly include PID control, fuzzy control, H¥ robust control, etc., and the longitudinal control methods mainly include sliding mode control, adaptive control, model predictive control, etc.


In summary, although the decoupling strategy has some effectiveness in vehicle formation control, the vehicle system is a non-linear, multi-variable, and strongly coupled system, in the conditions of higher acceleration, greater lateral and longitudinal forces, or lower road adhesion coefficients, the lateral and longitudinal coupling effects become particularly significant. During such conditions, the tracking performance of the decoupling control may decrease, leading to a significant reduction in control accuracy, which may lead to vehicle collisions and failure to achieve formation.


In addition, most of the current research on vehicle formation is based on the assumption of linear tires, which are only applicable within a limited operating range of the vehicle. When the vehicle moves out of the area (such as obstacle avoidance), the nonlinear and lateral and longitudinal coupled effects of the vehicle system become more pronounced, resulting in a mismatch between the decoupling control and emergency control of the vehicle.


SUMMARY

In light of these challenges raised in the above background technology, the purpose of the present application is to provide a reinforcement learning algorithm-based predictive control method for lateral and longitudinal coupled vehicle formation.


In order to achieve the aforementioned objectives, the present application provides the following technical solution:


A reinforcement learning algorithm-based predictive control method for lateral and longitudinal coupled vehicle formation, including:

    • S1, combining a 3-degree of freedom (DOF) vehicle dynamics model that takes into account a nonlinear magic formula tire model with a lane keeping model, in order to establish a vehicle formation model






x
i(k+1)=f(xi(k),ui(k));


in the formula, xi(k) is a state quantity and ui(k) is an input quantity;

    • S2, constructing a distributed control framework and designing a local predictive controller for each following vehicle based on the vehicle formation model under the control framework






J
i(xi(k),Ui(k))=Σl=0Tp-1∥xi(k+i)−ri(k+i)∥Qi2+∥xi(k+i)−{circumflex over (x)}i(k+i)∥Fi2+∥xi(k+i)−{circumflex over (x)}i-1(k+i)∥Gi2+∥ui(k+i)∥Ri2


in the formula, k is a current moment, k+i is a first moment in the prediction time domain, xi(•) is a prediction state, ri(•) is an ideal state, {circumflex over (x)}i-1(•) and {circumflex over (x)}i(•) represent assumed trajectory states of the vehicle, {circumflex over (x)}i-(•) is obtained through inter-vehicle communication, Tp is a prediction time domain, Qi, Fi, Gi, Ri are the weight matrices;

    • S3, using a reinforcement learning algorithm to solve the optimal control strategy of the local predictive controller and applying the optimal control strategy to the target following vehicle.


Preferably, the lateral force of the tire in the nonlinear magic formula tire model is calculated by the following magic formula:






F
i
y
=D sin(C arctan(Bα−E(Bα−arctan Bα)))

    • in the formula, α is a cornering angle of the tire, and B, C, D, E are simulation parameters.


Preferably, the 3-DOF vehicle dynamics model is expressed as:










?


(



m
i




v
.

i
x


-


m
i



v
i
y




φ
.

i



)


=

F
i
x













?


(



m
i




v
.

i
y


-


m
i



v
i
x




φ
.

i



)


=


F
i
yf

+

F
i
yr



;











?


I
i
z




φ
¨

i


=



a
i



F
i
yf


cos


δ
i


-


b
i



F
i

y

r












?

indicates text missing or illegible when filed




each parameter in the formula is the parameter of ith vehicle, and

    • vix,viy,{dot over (φ)}i are longitudinal speed, lateral speed, and yaw rate, Fix is a longitudinal force, Fiyf and Fiyr are the front and rear wheel lateral forces, mi is a vehicle mass, Iiz is a moment of inertia of the vehicle around the z axis, δi is the front wheel angle, ai and bi are the distances from the center of mass to the front and rear axles, respectively.


Preferably, the lane keeping model is expressed as:










?



e
.

i
p


=


v
i
x

-

v
0
x













?



e
.

i
y


=



v
i
x



e
i
φ


-

v
i
y

-

L



φ
.

i














?



e
.

i
φ


=




φ
.


i
,
des


-


φ
.

i










?

indicates text missing or illegible when filed






    • in the formula, {dot over (φ)}i,des is expected heading angular speed, L is a preview distance, eip is a longitudinal spacing error, eiy is a lateral position error between the vehicle and the lane line, and eiφ is a heading angle error between the vehicle heading angle and the road tangent.

    • combining the 3-DOF vehicle dynamics model with the lane keeping model to derive the vehicle formation model:













?



v
.

i
x


=



v
i
y




φ
.

i


+


1

m
i




F
i
x














?



v
.

i
y


=



-

v
i
x





φ
.

i


+



F
i

y

f


+

F
i

y

r




m
i















?



φ
¨

i


=


1

I
i
z




(



a
i



F
i

y

f



-


b
i



F
i

y

r




)



;











?



e
.

i
p


=


v
i
x

-

v
0
x













?



e
.

i
y


=



v
i
x



e
i
φ


-

v
i
y

-

L



φ
.

i














?



e
.

i
φ


=



φ
.


i
,
des


-


φ
.

i










?

indicates text missing or illegible when filed




By discretizing the model above and using the state variable xi=[vix viy {dot over (φ)}i eip eiy eiφ]T and the control variable ui=[Fix δi]T, xi(k+1)=f(xi(k),ui(k)) can be obtained.


Preferably, when utilizing the reinforcement learning algorithm to solve the optimal control strategy of the local predictive controller:

    • constructing and training an actor strategy function neural network to optimize the strategy parameters; specifically, when optimizing the strategy parameters, the actor strategy functional neural network uses a network composed of Tp radial basis functions to approximate the Tp-step optimal strategy, and the actor strategy functional neural network takes the state s as the input and the action a as the output;
    • constructing and training a critic value function neural network to evaluate the pros and cons of the current control strategy optimized by the actor strategy function neural network, specifically, when assessing the pros and cons of the current control strategy, the critic value function neural network is evaluated by a network consisting of Tp radial basis functions, and the critic value function neural network takes state s and action a as input and state-action value q(s,a) as output;
    • obtaining the optimal control strategy involves the alternating convergence of the actor strategy function neural network and the critic value function neural network.


Preferably, basis vectors ϕ(x) and ψ(x) in the actor strategy function neural network and the critic value function neural network are both radial basis functions, and










ϕ

(
x
)

=


ψ

(
x
)









=



(


exp


-




x
-

x
1




2


/

κ
2



,

exp


-




x
-

x
2




2


/

κ
2



,





e


-




x
-

x
M




2


/

κ
2





)

T


;











    • in the formula, κ is set to 1, {xi, i=1, 2, . . . M} as the center of the radial basis function, and M is a number of hidden layers.





Preferably, the center of radial basis function is obtained by normalization:








x
normalize

=



x
collect

-

x
min




x
max

-

x
min




;






    • in the formula, xnormalize is normalized data, xcollect is collected data, xmax and xmin are the maximum and minimum values in the collected data respectively;

    • the collected data involves a randomly given within the control input range, and the input data and output data of the vehicle formation model are collected by simulation.





Specifically, during the alternating convergence process to obtain the optimal control strategy:

    • initializing an actor strategy function neural network weight θ and a critic value function neural network weight ω;
    • obtaining action a by the actor strategy function neural network according to the current state s of the target following vehicle, and acting action a on the target following vehicle to obtain new state sc and instant reward r;
    • obtaining new action ac by the actor strategy function neural network according to the new state sc;
    • evaluating and scoring action a and action ac by the critic value function neural network to obtain q(s,a) and q(sii,a) and then calculating an error TDerror: TDerror=yi−q(s,a), yi=r+γq(sii,a) between a predicted value q(s,a) and an expected value yi of the critic value function neural network according to the Bellman equation;
    • using a gradient descent method to minimize the iterative update of the weight 9 and the weight ω, and the optimal control strategy Ui*, L(θ)=q(s,a), L(ω)=½TDerror2;


Preferably, in the predictive time domain, the optimal control strategy Ui* is applied to the target following vehicle through the local predictive controller.


Compared with the existing technology, the present application has the following beneficial effects:

    • (1) Compared with the traditional decoupling control method, the present application is more in line with the nonlinear, multi-variable, and strong coupling characteristics of the vehicle system, the lateral and longitudinal coupling modeling of the vehicle formation is completed, and the nonlinear characteristics of the tire are considered; In addition, under the framework of distributed model predictive control, a local predictive controller is designed for each following vehicle. This approach transforms the global optimization problem of vehicle formation into a local optimization problem for each following vehicle, which avoids the burden of centralized control.
    • (2) The reinforcement learning actor-critic algorithm is used to optimize the policy, specifically, the alternating critic training (policy evaluation) and actor training (policy improvement) are used to train the actor network to approach or approximate the optimal control strategy, compared with the common policy-based learning methods, the present application can realize the single-step update of the policy, thereby reducing the number of iterative learning.
    • (3) The present application uses the hidden layer space composed of radial basis functions to map high-dimensional input data thereby transforming the originally linear inseparable problem into a linearly separable one. This allows for the use of a linear optimization strategy to adjust the weights, resulting in a fast network learning speed and effectively improving the efficiency of solving optimization problems.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a schematic diagram of the 3-DOF vehicle dynamics model of the present application;



FIG. 2 is a Map table reflecting the relationship curve between the tire cornering angle and the tire lateral force.



FIG. 3 is a control flow chart of the local predictive controller of the present application;



FIG. 4 is a structure diagram of the reinforcement learning algorithm of the present application;



FIG. 5 is a network structure diagram composed of the radial basis function of the present application.





DETAILED DESCRIPTION OF THE EMBODIMENTS

The technical solution of the embodiment of the present application will be described clearly and completely, combined with the accompanying drawings. Obviously, the described embodiment is only part of the embodiments of the present application, not all of the embodiments. Based on the embodiment in the present application, all other embodiments obtained by ordinary technicians in this field without making creative labor belong to the scope of protection of the present application.


A reinforcement learning algorithm-based predictive control method for lateral and longitudinal coupled vehicle formation, including:


S1, combining the 3-DOF vehicle dynamics model that takes into account the nonlinear magic formula tire model with the lane keeping model, in order to establish the vehicle formation model;

    • among them:
    • as shown in FIG. 1, the 3-DOF vehicle dynamics model is expressed as










?


(



m
i




v
.

i
x


-


m
i



v
i
y




φ
.

i



)


=

F
i
x













?


(



m
i




v
.

i
y


-


m
i



v
i
x




φ
.

i



)


=


F
i
yf

+

F
i

y

r




;











?


I
i
z




φ
¨

i


=



a
i



F
i
yf


cos


δ
i


-


b
i



F
i

y

r












?

indicates text missing or illegible when filed






    •  each parameter in the formula is the parameter of ith vehicle, and

    • vix, viy, {dot over (φ)}i are longitudinal speed, lateral speed, and yaw rate, Fix is a longitudinal force, Fiyf and Fiyr are the front and rear wheel lateral forces, mi is a vehicle mass, Iiz is a moment of inertia of the vehicle around the z axis, δi is the front wheel angle, ai and bi are the distances from the center of mass to the front and rear axles, respectively.





The tire lateral force of the tire in the nonlinear magic formula tire model is calculated by the following magic formula:






F
i
y
=D sin(C arctan(Bα−E(Bα−arctan Bα)))

    • in the formula, α is a cornering angle of the tire, and B, C, D, E are simulation parameters.


Specifically, when considering the nonlinear magic formula tire model, calculating the longitudinal load of the front and rear tires first, and then obtaining the relationship curve between the tire sideslip angle and the tire lateral force by using the interpolation method with reference to FIG. 2, and then obtaining the parameter B, C, D, E by fitting the above magic formula of the tire.


The lane keeping model is expressed as.










?



e
.

i
p


=


v
i
x

-

v
0
x













?



e
.

i
y


=



v
i
x



e
i
φ


-

v
i
y

-

L



φ
.

i














?



e
.

i
φ


=



φ
.


i
,
des


-


φ
.

i










?

indicates text missing or illegible when filed




in the formula, {dot over (φ)}i,des is expected heading angular speed, L is a preview distance, eip is a longitudinal spacing error, eiy is a lateral position error between the vehicle and the lane line, and eiφ is a heading angle error between the vehicle heading angle and the road tangent.


In summary, combining the above 3-DOF vehicle dynamics model with the lane keeping model, the vehicle formation model can be obtained as follows;










?



v
.

i
x


=



v
i
y




φ
.

i


+


1

m
i




F
i
x














?



v
.

i
y


=



-

v
i
x





φ
.

i


+



F
i

y

f


+

F
i

y

r




m
i















?



φ
¨

i


=


1

I
i
z




(



a
i



F
i

y

f



-


b
i



F
i

y

r




)



;











?



e
.

i
p


=


v
i
x

-

v
0
x













?



e
.

i
y


=



v
i
x



e
i
φ


-

v
i
y

-

L



φ
.

i














?



e
.

i
φ


=



φ
.


i
,
des


-


φ
.

i










?

indicates text missing or illegible when filed






    • taking the state quantity xi=[vix viy {dot over (φ)}i eip eiy eiφ]T, the control quantity ui=[Fix δi]T and the sampling time Ts, the discrete form of the vehicle formation model can be obtained after discretization of the above vehicle formation model: xi(k+1)=f(xi(k),ui(k)); in the formula, xi(k) is the state variable and ui(k) is the input variable.

    • S2, constructing a distributed control framework and designing a local predictive controller for each following vehicle based on the vehicle formation model under the control framework









J
i(xi(k),Ui(k))=Σl=0Tp-1∥xi(k+i)−ri(k+i)∥Qi2+∥xi(k+i)−{circumflex over (x)}i(k+i)∥Fi2+∥xi(k+i)−{circumflex over (x)}i-1(k+i)∥Gi2+∥ui(k+i)∥Ri2


in the formula, k is a current moment, k+i is a first moment in the prediction time domain, xi(•) is a prediction state, ri(•) is an ideal state, {circumflex over (x)}i-1(•) and {circumflex over (x)}i(•) represent assumed trajectory states of the vehicle, {circumflex over (x)}i-(•) is obtained through inter-vehicle communication, Tp is a prediction time domain, Qi, Fi, Gi, Ri are the weight matrices; where the assumed input of each vehicle is defined as follows:












u
^

i

(


k
+
j

|

k
+
1


)

=






?



u
i
*

(


k
+
j

|
k

)


,

j
=
0

,






T
p


-
2









?

0

,

j
=


T
p

-
1







;








?

indicates text missing or illegible when filed




It is assumed that the trajectory can be calculated from the assumed input:










?




x
^

i

(


k
+
j
+
1

|

k
+
1


)


=

f

(




x
^

i

(


k
+
j

|

k
+
1


)

,



u
^

i

(


k
+
j

|

k
+
1


)


)












?




x
^

i

(


k
+
1

|

k
+
1


)


=


x
i
*

(


k
+
1

|
k

)









?

indicates text missing or illegible when filed






    • S3, using a reinforcement learning algorithm to solve the optimal control strategy of the local predictive controller and applying the optimal control strategy to the target following vehicle through the local predictive controller.





(31) Constructing the actor strategy function neural network and the critic value function neural network.


Specifically, combined with the structure shown in FIG. 4, the actor strategy function neural network and the critic value function neural network are set as follows:


The actor strategy function neural network uses a network consisting of Tp radial basis functions to approximate the Tp-step optimal strategy;

    • evaluating the critical value function neural network by a network consisting of Tp radial basis functions;
    • the network structure composed of Tp radial basis functions is shown in FIG. 5, in the approximation optimal strategy, the actor strategy function neural network takes state s as input and action a as output; in the evaluation, the critic value function neural network takes state s and action a as input and state-action value q(s,a) as output.


Preferably, the basis vectors ϕ(x) and ψ(x) in the actor strategy function neural network and the critic value function neural network are radial basis functions, and










ϕ

(
x
)

=


ψ

(
x
)







=



(


exp


-




x
-

x
1




2


/

κ
2



,

exp


-




x
-

x
2




2


/

κ
2



,





e


-




x
-

x
M




2


/

κ
2





)

T








In the formula, x is set to 1, {xi=1, 2, . . . M} as the center of the radial basis function, and M is the number of hidden layers.


The center of radial basis function is obtained by normalization:








x
normalize

=



x
collect

-

x
min




x
max

-

x
min




;






    • in the formula, xnormalize is normalized data, xcollect is collected data, xmax and xmin are the maximum and minimum values in the collected data respectively; specifically, the collected data involves a randomly given within the control input range, and the input data and output data of the vehicle formation model are collected by simulation.





(32) Training actor strategy function neural network and critic value function neural network.


initializing the actor strategy function neural network weight θ and the critic value function neural network weight ω;


obtaining action a by the actor strategy function neural network according to the current state s of the target following vehicle, and acting action a on the target following vehicle to obtain new state sc and instant reward r;

    • obtaining new action ac by the actor strategy function neural network according to the new state sc;
    • evaluating and scoring action a and action ac by the critic value function neural network to obtain q(s,a) and q(sii,a) and then calculating an error TDerror: TDerror=yi−q(s,a), yi=r+γq(sii,a) between a predicted value q(s,a) and an expected value yi of the critic value function neural network according to the Bellman equation;


in order to minimize the value function obtained by the action output of the actor strategy function neural network, taking the value function q(s,a) as the loss function L(θ) of the actor strategy function neural network, that is, L(θ)=q(s,a), and using the gradient descent method to iteratively update the weight θ; in order to make the score of the critic value function neural network more accurate, taking the loss function L(ω)=½TDerror2 of the critic value function neural network, and updating the weight ω iteratively by the gradient descent method; specifically, when the number of iterations or the accuracy meets the preset conditions, the optimal control strategy Ui* is obtained.


(33) Using the above actor strategy function neural network and critic value function neural network to solve the optimal control strategy of the local predictive controller, and the optimal control strategy Ui* solved in the prediction time domain acts on the target following vehicle through the local predictive controller.


Although the embodiment of the present application has been presented and described, it is understandable to ordinary technicians in the field that these embodiments can be varied, modified, replaced, and amended without departing from the principles and spirit of the present application, Therefore, the scope of the present application is limited by the accompanying claims and their equivalents.

Claims
  • 1. A reinforcement learning algorithm-based predictive control method for a lateral and longitudinal coupled vehicle formation, comprising: S1, combining a 3-degree of freedom (DOF) vehicle dynamics model with a lane keeping model to establish a vehicle formation model, wherein the 3-DOF vehicle dynamics model takes into account a nonlinear magic formula tire model; xi(k+1)=f(xi(k),ui(k));
  • 2. The reinforcement learning algorithm-based predictive control method for the lateral and longitudinal coupled vehicle formation according to claim 1, wherein a lateral force of a tire in the nonlinear magic formula tire model is calculated by the following magic formula: Fiy=D sin(C arctan(Bα−E(Bα−arctan Bα)))wherein α is a cornering angle of the tire, and B, C, D, E are simulation parameters.
  • 3. The reinforcement learning algorithm-based predictive control method for the lateral and longitudinal coupled vehicle formation according to claim 1, wherein the 3-DOF vehicle dynamics model is expressed by the following formula:
  • 4. The reinforcement learning algorithm-based predictive control method for the lateral and longitudinal coupled vehicle formation according to claim 3, wherein the lane keeping model is expressed as:
  • 5. The reinforcement learning algorithm-based predictive control method for the lateral and longitudinal coupled vehicle formation according to claim 1, wherein when utilizing the reinforcement learning algorithm to solve the optimal control strategy of the local predictive controller: constructing and training an actor strategy function neural network to optimize strategy parameters;constructing and training a critic value function neural network to evaluate a pros and a cons of a current control strategy optimized by the actor strategy function neural network;obtaining the optimal control strategy according to an alternating convergence of the actor strategy function neural network and the critic value function neural network.
  • 6. The reinforcement learning algorithm-based predictive control method for the lateral and longitudinal coupled vehicle formation according to claim 5, wherein when optimizing the strategy parameters, the actor strategy functional neural network uses a network composed of Tp radial basis functions to approximate a Tp-step optimal strategy and takes a state s as a first input and an action a as a first output; when assessing the pros and the cons of the current control strategy, the critic value function neural network is evaluated by the network composed of the Tp radial basis functions and takes the state s and the action a as a second input and a predicted value q(s,a) as a second output.
  • 7. The reinforcement learning algorithm-based predictive control method for the lateral and longitudinal coupled vehicle formation according to claim 6, wherein basis vectors ϕ(x) and ψ(x) in the actor strategy function neural network and the critic value function neural network are the radial basis functions, and
  • 8. The reinforcement learning algorithm-based predictive control method for the lateral and longitudinal coupled vehicle formation according to claim 7, wherein the center of the radial basis functions is obtained by a normalization:
  • 9. The reinforcement learning algorithm-based predictive control method for the lateral and longitudinal coupled vehicle formation according to claim 8, wherein the optimal control strategy is obtained through the alternating convergence; initializing an actor strategy function neural network weight 9 and a critic value function neural network weight ω;obtaining the action a by the actor strategy function neural network according to a current state s of the each target following vehicle, and acting the action a on the each target following vehicle to obtain a state sc and an instant reward r;obtaining an action ac by the actor strategy function neural network according to the state sc;evaluating and scoring the action a and the action ac by the critic value function neural network to obtain the predicted value q(s,a) and a predicted value q(sii,a) and then calculating an error TDerror: TDerror=yi−q(s,a), yi=r+γq(sii,a) between the predicted value q(s,a) and an expected value yi of the critic value function neural network according to a Bellman equation;using a gradient descent method to minimize an iterative update of the actor strategy function neural network weight θ and the critic value function neural network weight ω to obtain the optimal control strategy Ui*, L(θ)=q(s,a), L(ω)=½TDerror2.
  • 10. The reinforcement learning algorithm-based predictive control method for the lateral and longitudinal coupled vehicle formation according to claim 8, wherein in the prediction time domain, the optimal control strategy Ui* is applied to the each target following vehicle through the local predictive controller.
Priority Claims (1)
Number Date Country Kind
202211087600.7 Sep 2022 CN national