This application claims the benefit of priority from Chinese Patent Application No. 202410006216.2, filed on Jan. 3, 2024. The content of the aforementioned application, including any intervening amendments made thereto, is incorporated herein by reference in its entirety.
This application relates to shoreline segmentation, and more particularly to a method and device for shoreline segmentation in complex environments based on the perspective of an unmanned surface vessel (USV).
There is a lack of an unmanned surface vessel (USV)-based shoreline recognition and segmentation system suitable for dynamic and complex aquatic environments in the prior art. Current systems mainly focus on the recognition in static environments, without fully considering the environmental diversity and the self-control of USV. Furthermore, these systems generally rely on the data from a single sensor, require high lighting conditions, and cannot achieve all-weather real-time recognition. Due to the size constraints of the USVs and the limitation of the computational power of relevant equipment, it is difficult to apply high-performance models for the practical inspection, thereby hindering further development.
Chinese Patent Application No. 202211018479.2, titled “Method for constructing water surface environment map based on millimeter wave radar”, employs a millimeter wave radar and visible light images to extract shoreline information. However, this method does not involve the vessel's self-control, and fails to completely extract the shoreline depth features. Moreover, this method has poor robustness due to reliance on single visible light images.
Chinese Patent Application No. 201410764435.3, titled “Shoreline registration method based on radar images and electronic chart information”, discloses a shoreline registration method based on radar images and electronic chart information, in which radar charts are processed and registered with accurate electronic charts to achieve the precise shoreline extraction. However, it does not address the shoreline extraction in the absence of accurate charts, and does not involve the vessel self-control.
Chinese Patent Application No. 202310790874.0, titled “Multi-sensor situation awareness and collision avoidance decision-making method for unmanned surface vessels”, designs a collision avoidance perception system for the unmanned surface vessel based on visible light images and radar, and adopts the artificial potential field method for path planning. However, this method has poor robustness when the visible light images are captured in complex environments, and serious deficiencies of single radar features, Moreover, the artificial potential field method fails to adapt to rapidly changing complex environments.
Chinese Patent Application No. 202010717418.X, titled “Unmanned surface vessel path planning method based on deep reinforcement learning and marine environmental elements”, designs a deep reinforcement learning path planning algorithm related to marine factors, but does not involve the multi-source perception to the environment, nor does it provide a self-control scheme for the vessel.
The USV control can be achieved through the separation of longitudinal and steering motions in combination with adaptive control and an adaptive line-of-sight (LOS) guidance algorithm for trajectory tracking (M. Faramin, R. H. Goudarzi, A. Maleki, Track-keeping Observer-based Robust Adaptive Control of an Unmanned Surface Vessel by Applying a 4-DOF Maneuvering Model, Ocean Engineering, Volume 183, 2019, Pages 11-23). Although this method is relatively robust, it does not reserve interfaces for other tasks, and has weak scalability, and poor adaptability to complex situations.
In view of the problems in the prior art, this application provides a method and device for shoreline segmentation in complex environments based on the perspective of an unmanned surface vessel.
The present disclosure provides a method for shoreline segmentation in complex environments based on the perspective of an unmanned surface vessel, comprising:
In an embodiment, the fusion and feasible region segmentation are performed by using a dual modal segmentation network (DMSNet), wherein the DMSNet comprises an encoder, a decoder, and a dual-path feature spatial adaptation (DPFSA) module; and
In an embodiment, the step of obtaining the echo image based on the raw radar echo image comprises:
In an embodiment, the step (S3) comprises:
In an embodiment, the step (S32) comprises:
In an embodiment, the boundary cost β is calculated according to the following formula:
In an embodiment, the method comprising:
In an embodiment, the step (E) further comprises:
b, a1, a2 and a3 represent control parameters; Sd=u−ud, Sd represents a difference between the forward speed and the desired longitudinal speed; ud represents a desired longitudinal speed; μ represents an adaptation rate parameter, and is configured to control an update rate of a system parameter; α is a speed weight parameter and configured to weigh system performance; μ and α are design parameters; and p represents a roll rate of the USV.
In an embodiment, the new state of the USV is estimated by the observer according to the following formula:
The present disclosure further provides a shoreline segmentation device, comprising
By adjusting the feedback gain, the system's response speed and accuracy can be improved. In adaptive control, the adjustment of the feedback gain helps the system adapt more quickly to changing environments or operating conditions. When facing external disturbances or internal parameter changes, the adjustment of the feedback gain can enhance the system's robustness, ensuring good performance even under sub-optimal conditions. Proper adjustment of the feedback gain can reduce system overshoot and oscillation, thereby improving response quality. The adjustment of the feedback gain is a continuous process that dynamically adapts to changes in the system state or environment.
As an inventive concept, this present application also provides a terminal device, which includes a memory, a processor, and a computer program stored in the memory. The processor executes the computer program stored in the memory to implement the steps of the method described in this present application.
Compared with the prior art, the beneficial effects of this present application are as follows. The present application focuses on the effective segmentation and extraction of shoreline information in complex environments. To address issues of weak robustness, limited tiny features of the shoreline and limited depth information of the image captured under a single light source in complex environments, a stacked ensemble shoreline segmentation model is proposed. Further, to address the issue of underactuated USV in complex environments being unable to effectively navigate along a preset route, which results in the shoreline segmentation model failing to achieve its intended results, a control module for distance to the shoreline has been added. A deep reinforcement learning-based marine path planning model is proposed, and commands from this model are converted into USV actions by the demand-controlled ventilation (DCV) robust adaptive control unit, maintaining the USV's distance to the shoreline, an appropriate speed, and timely obstacle avoidance, thereby ensuring the shoreline segmentation model achieves the expected results.
The present disclosure has been described in detail above with reference to accompanying drawings and embodiments. It should be noted that the disclosed embodiments are merely exemplary, and are not limited to limit the present disclosure. Those skilled in the art can still make various changes, modifications and replacements to technical features recited in the above embodiments. It should be understood that those changes, modifications and replacements made without departing from the spirit of the disclosure shall fall within the scope of the disclosure defined by the appended claims.
To achieve shoreline segmentation in complex environments based on the perspective of the USV, embodiment 1 of the present application is illustrated in two parts: a stacking ensemble-based shoreline segmentation model and a distance-to-shoreline control module based on deep reinforcement learning and adaptive robust control. The sensors used herein include a red-green-blue (RGB) camera, a thermal infrared camera, a short-wave pulse radar, and LiDAR, as illustrated in
The specific composition and workflow are as follows.
The shoreline segmentation model consists of three sub-modules: dual-light fusion and feasible region segmentation, shoreline boundary extraction by short-wave pulse radar, and shoreline point cloud extraction by LiDAR. Visible light images and thermal infrared images are captured using an RGB camera and a thermal infrared camera, respectively. The visible light image and the thermal infrared image are subjected to fusion and feasible region segmentation to obtain all-weather two-dimensional image information of the shoreline. Short-wave pulse radar is used to obtain the echo image containing tiny features. The characteristic signal intensity of the tiny features is less than three times the background noise of the tiny features. Both the all-weather two-dimensional image information and the echo image are input into the process of shoreline point cloud extraction by LiDAR to constrain the extraction region and enhance shoreline features. The final output is a multi-feature point cloud dataset for shoreline segmentation.
A USV feasible region segmentation model, DMSNet, is employed for visible and thermal infrared image fusion from the perspective of the USV. Extracting feasible regions within the surrounding marine environment is fundamental to achieving autonomous positioning and navigation for the USV on the water surface. Addressing the challenge of all-weather shoreline perception in dark environments, a DMSNet is proposed for the USV under dark environments, based on a visible light camera and an infrared thermal imaging device. The flowchart of the model is shown in
DMSNet primarily consists of an encoder and a decoder. The encoder uses two separate paths to extract features from visible light and thermal infrared images. Each path includes five sets of operations, with each set containing one to three 3×3 convolutional layers followed by batch normalization (BN) layers, which help maintain stable feature distribution within the network. Between each set, a max pooling layer with a stride of 2 reduces the spatial dimensions of the feature maps while increasing the number of convolutional kernels, allowing the encoder to progressively learn richer semantic information from shallow to deep layers. Since DMSNet is a lightweight network designed for the perception of the marine environment, it uses a leaky rectified linear unit (Leaky-ReLU) as the activation function throughout the network. The final output is a segmentation result with the same dimensions as the input image. In the decoder, each set of operations is similar to that of the encoder and contains convolutional layers, batch normalization layers, and activation functions. Between sets of the decoder, a fast upsampling method with a scaling factor of 2 using nearest-neighbor interpolation is applied. Additionally, feature maps of the same dimensions from both the visible light and thermal infrared encoders are fused. To reduce the differences in feature space between the two modalities, the DPFSA module is introduced to automatically transform both modalities into the same space for fusion.
The DPFSA module is designed to perform spatial feature extraction. It retains the feature vectors of data from different modalities and includes pre-adaptation operations and inverse transformation layers. The pre-adaptation operations enhance the model's non-linear capabilities, further applying convolution operations to the transformed data to prevent significant changes in data distribution and improve the model's flexibility. These enhancements allow the shoreline segmentation model to improve its segmentation performance with minimal increase in network parameters.
This module primarily includes two functions: spatial feature extraction (transformation of feature spaces) and fusion of features carrying different information. For spatial feature extraction, a 1×1 convolutional layer with Leaky-ReLU is first used to pre-adapt the thermal infrared features. The pre-adapted thermal infrared features and the visible light features are then input into the transformation network (TransNet) to learn the transformation parameters. Finally, the inverse transformation layer completes the space transformation of the thermal infrared features, as shown in formula (1):
In the above formula, α=Hα(fpre_ther, fvis; Wα) represents the transformed thermal infrared features; α=Hα(fpre_ther, fvis; Wα) represents the operation of the inverse transformation layer, which has the same structure as the pre-adaptation; α=Hα(fpre_ther, fvis; Wα) represents the pre-adapted thermal infrared features; α=Hα(fpre_ther, fvis; Wα) and α=Hα(fpre_ther, fvis; Wα) represent the transformation parameters output by TransNet, calculated by two sub-networks within TransNet as follows:
In the above formulas, Hα and Hβ represent the full convolution operations for α and β calculated by the two transformation sub-networks; Wα and Wβ are the corresponding parameters; and fvis represents the visible light features.
After completing the transformation of feature spaces, the next step is to fuse the features. The transformed thermal infrared features are first concatenated with the visible light features, then element-wise added with the previously fused results to achieve a fusion effect, forming dual-path features. For the first set of data, the initial fusion features are directly obtained by using their own dual-light features (the fusion result of the thermal infrared and visible light features). The process handled by the DPFSA module can be represented as shown in formula (4):
In this above formula, n is the number of groups in the scene segmentation model; V and T represent the visible light image and the thermal infrared image, respectively; ffuse represents the result after a group of convolution operations in the decoder for the dual-path features output by the DPFSA module; W represents all parameters of this module; and Mfuse represents the fusion process through point-by-point addition. As seen in formula (4), the DPFSA module not only retains information from both modalities to form dual-path features but also allows these dual-path features to be processed and then input into the next DPFSA module. This method maximally reduces information mixing and loss, enhancing the utilization of thermal infrared images.
Infrared images from thermal imaging typically have low contrast, with blurred target contours, making it easy for targets to blend with the background and cause misidentifications. By optimizing data augmentation techniques, including horizontal/vertical flipping, rotation/reflection, and random scaling, and by applying contrast-limited histogram equalization for detail enhancement, clearer boundaries and sharper contours of obstacles are achieved. Additionally, a new dataset label format has been introduced, containing directional connectivity information between pixels of the same category. Each category can form an eight-channel directional label map for training, enabling the network to learn the connectivity relationships among similar pixels. This helps mitigate the inaccuracies caused by low contrast and blurred target contours in infrared images, thereby enhancing model prediction accuracy. The output consists of image data, with echo image being a dataset made up of pixels, where each pixel can be represented as q=[xq
In the process of extracting and parameterizing shoreline features sensed by the short-wave pulse radar, morphological and bilateral filters are applied to enhance the quality of shoreline images from radar images that commonly contain noise and clutter in ocean environments. A median filter is used to reduce outliers and artifacts in the obtained shoreline features, as illustrated in
In the above formula, p represents the position vector of the spline curve; ci represents the control points; Bik represents the normalized basis functions defined by the Cox-de Boor recursion formula; the equation of feature points in a B-spline curve can be expressed as shown in formula (6):
In the above formula, pj represents the j-th feature point obtained from the radar image; tj represents the position along the curve and can be approximated as follows:
Given multiple feature points P, a set of control points C can be obtained by fitting these feature points. The least squares fitting equation is expressed as follows:
In the above formula, B represents a configuration matrix with a maximum of k non-zero values per row. A cubic spline curve (k=4) represents the coastline curve, applying a clamped knot vector so that the ends of the spline curve coincide with the first and last control points. An example of coastline parameters obtained using radar images is shown in
The point cloud p obtained from the LiDAR sensor dataset is represented as P=[xe
The process begins with the detection of candidate boundary points. During initialization, all point clouds are considered as unmarked points. As shown in
During candidate point extraction, missing points may occur, which is related to the chosen value of neighboring points k. A smaller k can produce more complete boundary points but requires more time for candidate point extraction. Additionally, if a candidate point is too far from other candidate points, exceeding a set threshold Td, it is considered an outlier and will be removed. Due to the point cloud's non-uniform and unordered nature, extracting all boundary points along the shoreline is challenging, because candidate points can generate different boundaries. Assuming there are five candidate boundary points, as shown in the upper left corner of
Then, the boundary cost is calculated using the energy function. To evaluate the quality of the shoreline point cloud boundary, the boundary cost β can be defined as follows:
In the above formula, n represents the number of connections in the boundary, λ is a weight coefficient, Bi represents an i-th connection line, D(Bi) is the length of the i-th connection Bi, and <Bi,Bj> represents the angle between the adjacent connection lines. During boundary optimization, it is desirable for boundary points to be close to each other (D(Bi) is small) and for the angle between adjacent connection lines to be large (cos(<Bi,Bj>/2) is close to 0).
Finally, the boundary is optimized by obtaining the minimum-cost boundary. To obtain a boundary with the minimum β value, a minimum-cost boundary model is used to select the optimal shoreline boundary, as shown in
represents the minimization objective function that requires that the path starts from the Source, passes through a series of nodes, and finally reaches the Sink. Node {IJK} is treated as equivalent to {KJI}, reducing the computational complexity. Thus, the combination in the upper right corner of
The shoreline point cloud data E is obtained from q, g, and p, and is represented as E=(xe
The control module for distance to the shoreline is configured to stabilize the speed and distance of the USV from the shoreline in a complex environment, providing an optimal working environment for sensors. It is divided into two steps: deep reinforcement learning-based marine path planning and DCV robust adaptive control. Firstly, the deep reinforcement learning-based marine path planning part gathers distance information between the USV, the shoreline, and obstacles from segmented shoreline point cloud data. The USV's real-time speed and heading are obtained from onboard feedback. By rewarding the maintenance of a preset speed and distance to the shoreline, the system outputs appropriate values for speed and heading adjustments. To implement these route planning instructions on the USV, a DCV robust adaptive controller transforms them into force and torque adjustments for the USV's power system.
The input data includes the shoreline point cloud data E and the data S collected by the sensors onboard the USV, which encompass its current speed, heading, distance to obstacles, ocean current vector speed, and wind vector speed. Through E and S, information such as the distance of the USV to obstacles, its relative distance to the shoreline, current heading, and speed can be obtained.
S=(xs
The distance between the USV and the closest shoreline point can be calculated from E and S as follows:
dUSV-to-Shore represents the distance from USV to the shoreline, the function (mini) is configured to find the minimum Euclidean distance between the USV's position and any point on the shoreline.
In summary, the observation state d can be expressed as:
The actions performed by the USV in the sea area are represented as a=[wl,ud], where w1 represents the heading adjustment value, indicating the USV's heading change in degrees (positive for right turns, negative for left turns); ud represents the reference speed, representing the target speed for the USV in the upcoming period, measured in knots (positive for acceleration, negative for deceleration). Actions are generated by the agent network and converted into point cloud data representation through a decoder.
The reward function is configured to guide the model in learning to make correct navigation decisions, achieving goals like avoiding surrounding obstacles, maintaining an ideal distance from the coastline, and traveling along the optimal path. In the marine path planning based on deep reinforcement learning, the intelligent agent's objectives and motivations (e.g., obstacle avoidance and following the best path) are defined.
When the USV successfully avoids surrounding obstacles and moves toward the target direction, it receives a positive reward. If the USV collides with an obstacle or deviates from the path, it receives a negative reward. The reward function is as follows:
In the above formula, angle represents the angle between the USV's travel direction and the tangent of the planned path, and trackPos represents the normalized distance from the USV's centroid to the planned path.
3D CNN is configured for processing point cloud data and extracting spatial features, and includes the following layers:
The loss function is configured to measure the difference between the model's predictions (predicted action values) and actual observations, guiding the optimization of the path-planning model to better learn and predict which actions will yield the best results.
The overall loss function is the weighted sum of three individual loss functions, as expressed:
In the above formula, λactor, λcritic, and λdecoder represent weights used to balance the contributions of each loss component.
Actor Network Loss Function is represented by:
In the above formula, Lactor represents the loss function of the actor network; E[·] represents the expected value, indicating the average over all possible actions; log(π(a/s)) represents the natural logarithm of the policy probability of action a in state s; π(a/s) represents the policy function, giving the probability of selecting action a in state s; R(s,a) represents the reward obtained by taking action a in state s. This loss function encourages the actor network to increase the probability of high-reward actions and decrease the probability of low-reward actions. A smaller value indicates that the predicted actions are more effective.
Critic Network Loss Function is represented by:
Lcritic represents the loss function of the critic network; Qpredicted (s,a) represents the value of taking action a in state s as predicted by the critic network; Qtarget(s,a) represents the target value, calculated based on the Bellman equation; E[(·)2] represents mean squared error, measuring the gap between the predicted and target values. This loss function optimizes the critic network's ability to estimate action values accurately. A smaller value indicates that the critic network is more precise in estimating action values.
The decoder loss function is represented by:
In the above formula, Ldecoder represents the loss function of the decoder; decoder_output represents the output of the decoder (the converted action or control command); target_command represents the expected control command or action; E[∥19 ∥2] represents mean squared error, measuring the difference between the decoder's output and the expected command. This loss function ensures that the decoder can accurately convert the actor network's output into executable control commands. By minimizing this loss, the decoder learns to better interpret and convert the actor network's decisions, with a smaller value indicating more accurate conversions.
Step 1: the USV's state data S is obtained, and the shoreline point cloud data E processed from the shoreline segmentation model is received. Based on S and E, the USV's observation state d and reward function value r are determined. Step 2: d and r are input into the deep reinforcement learning-based marine path planning model. This model, using its internal structure (including 3D CNN and AC network), encodes the data, predicts the optimal navigation actions, and evaluates the predicted best navigation actions. The optimal actions include adjustments in direction and changes in speed. Step 3: the predicted optimal navigation actions are converted into specific instructions that the USV can understand and execute (action a), via the decoder. Step 4: Based on the action a executed by USV, the data in the state space is updated and a new observation state dl and reward function value r1 for the USV are provided. This information is sent back to the path planning model, guiding it in adjusting the prediction of optimal navigation actions.
Through this workflow, the USV maintains an appropriate distance from the shoreline and a certain navigation speed, performing obstacle avoidance maneuvers when necessary.
Environmental disturbances such as wind direction and speed, as well as ocean currents and flow speed, are used as inputs along with heading and speed changes from the path planning module, producing outputs for force and moment changes.
The navigation data is obtained using data S and action a, involving variables such as vs
A sliding mode controller is used to respond to course errors of USV by generating corrective actions to reduce deviation. The controller incorporates environmental disturbance estimations and adjusts the USV's power system to counteract interferences like wind and currents. The adaptive sliding mode controller is formulated as follows:
In the above formula, {dot over (u)} represents a longitudinal acceleration of the USV, where a longitudinal direction is a heading direction of the USV; m represents a weight of the USV; Xu, Y{dot over (v)}, Y{dot over (r)}, Y{dot over (p)} represent hydrodynamic coefficients related to motion of the USV; v represents a lateral speed of the USV; r represents a yaw rate of the USV; lg represents a longitudinal position of a centroid of the USV; hg represents a vertical position of the centroid of the USV; R(u) represents a hydrodynamic resistance of the USV at a forward speed of u; τcu represents a thrust force in the input command; D1(t) represents an environment-related disturbance term and is, derived from the longitudinal components of ccurrent
The control input with disturbance rejection is as follows:
In the above formula, b, a1, a2 and a3 represent control parameters; Sd=u−ud;Sd represents a difference between the forward speed and the desired longitudinal speed; ud represents a desired longitudinal speed; μrepresents an adaptation rate parameter, and is configured to control an update rate of a system parameter; α is a speed weight parameter and configured to weigh system performance; μ and α are design parameters; and p represents a roll rate of the USV.
Low-level control allocation is achieved by separating different dynamic characteristics of the USV (such as surge and yaw) and applying specialized control strategies for each. A 4-DOF dynamic maneuvering model is used, which includes surge, sway, yaw, and roll motions. These motions are managed for USV control, particularly under wind disturbances. In this model, surge motion is decoupled from the other degrees of freedom, with adaptive sliding mode control used as a robust controller to estimate unknown parameters and manage the desired longitudinal speed. Additionally, a robust adaptive state feedback controller with an integral term of the output tracking error is employed as a new state variable to track the reference heading angle.
The propulsion system of the USV is adjusted according to the controller output, which involves modifying thrust magnitude and thrust direction. These adjustments are fed back into the USV's propulsion and rudder systems to achieve precise control over heading and speed. Based on the USV's dynamic model and control commands, heading and speed adjustments are executed. The dynamic model includes factors such as the USV's weight, thrust, and rudder angle.
In the above formula, η represents position and orientation vector, indicating the USV's position and orientation in the Earth's coordinate system; vs
Related dynamic equations are represented by:
In the above formula, the first equation of formula (19) describes the USV's linear dynamics in the direction of its course; m represents the weight of the USV; X{dot over (u)}, Y{dot over (v)}, Y{dot over (p)} are coefficients related to motion; v, r, p represent speed components; R(u) represents hydrodynamic drag function; τcu, τdu represent control input terms, indicating the forces along the body's X-axis (longitudinal), affecting the USV's surge motion and the external disturbances respectively. The second equation of formula (19) describes the USV's dynamics in the vertical and rotational directions. Similarly, m presents the weight of the USV; Y{dot over (v)}, Y{dot over (r)}, are coefficients related to motion; v, r, p represent speed components; Yv, Yp, Yr, Yϕ represent force and torque coefficients; u represents forward speed; τcv, τdv represent control input terms, indicating forces along the body's Y-axis (lateral), influencing the USV's sway motion and external disturbances, respectively. The third equation of formula (19) describes the roll dynamics of the USV. Ixx represents roll inertia of the USV; mlg, mhg, N{dot over (v)}, K{dot over (p)} are coefficients related to motion; Kv, Kp, Kr, Kϕ represent force and torque coefficients; mGMT represents static stability moment of the roll; zg represents height of the USV's center of gravity; u represents forward speed. τcp, τdp represent control input terms, indicating torques, which affect roll and external disturbances, respectively. The fourth equation of the formula (19) describes the yaw dynamics of the USV; lzz represents yaw inertia of the USV; N{dot over (r)} is Coefficient related to motion; Nv, Np, Nr, Nϕ are force and torque coefficients; m1g is a coefficient related to weight and geometry parameters; u represents forward speed; τcr, τdr are control input terms, representing torques, which affect yaw and external disturbances, respectively.
This step generates control commands for the USV's propulsion and rudder systems. These commands are applied to respond to the planned course and environmental disturbances, maintaining the USV's navigation along the predetermined path.
The course controller uses the current environmental conditions as disturbances and lateral thrust derived from the desired heading angle change and the sliding mode controller's output as rudder control input to calculate the rudder angle change yp to adjust the USV's heading:
In the above formula, xp represents New state vector, including speed and heading angle; τcv represents rudder control input, indicating force along the body's Y-axis axis (lateral), affecting the USV's sway motion, calculated as
with the angle provided by a; d2(t) represents environmental disturbance, provided by the lateral components ccurrent
An integrated output tracking error vector calculates the difference between the current and target states, indicating the output tracking error. This error vector reflects the deviation between the actual performance of the USV and its preset targets (e.g., heading angle, position) and is calculated as:
In the above formula, eyl represents integrated output tracking error.
ycmd represents a commanded heading angle, Cp represents a transformation matrix.
An observer estimates the USV's new state based on its dynamic model and control inputs:
In the above formula, {circumflex over (x)} represents an estimated state vector; A represents system matrix, indicating the USV's intrinsic dynamic behavior in the state without control input; B represents input matrix, describing an influence of the forward speed u; L represents a gain matrix of observer to ensure that the estimated state from the the USV's dynamic system converges to the actual state; C represents output matrix, describing a transformation of the state of the dynamic system to an output value; and K represents feedback gain.
The adjustment of the controller gain is performed according to the output tracking error to reduce the error and enhance controller performance:
In the above formula, {dot over (K)} represents an adjustment change rate of the feedback gain; Γ represents a parameter for adjusting a response rate of the adaptive law; and P1 represents a positive definite matrix related to a state estimation error.
This Embodiment 2 of the present disclosure provides a terminal device corresponding to Embodiment 1 above. The terminal device can be a processing device for a client, such as a mobile phone, laptop, tablet, desktop computer, etc., to execute the method in Embodiment 1.
The terminal device in this embodiment includes a memory, a processor, and a computer program stored in the memory. The processor executes the computer program stored in the memory to implement the steps of the method in Embodiment 1. In some implementations, the memory can be high-speed random access memory (RAM) and may also include non-volatile memory, such as at least one disk storage. In other implementations, the processor can be various types of general-purpose processors, such as a central processing unit (CPU) or a digital signal processor (DSP), without limitation here.
This Embodiment 3 of the present disclosure provides a computer-readable storage medium corresponding to Embodiment 1 above, on which a computer program/instructions are stored. When executed by the processor, the computer program/instructions implement the steps of the method in Embodiment 1.
The computer-readable storage medium can be a tangible device that retains and stores instructions used by an instruction execution device. For example, the computer-readable storage medium may include, but is not limited to, electrical storage devices, magnetic storage devices, optical storage devices, electromagnetic storage devices, semiconductor storage devices, or any combination of the above.
Those skilled in the art should understand that the embodiments of this application can be provided as a method, a system, or a computer program product. Therefore, this application can take the form of a fully hardware-based embodiment, a fully software-based embodiment, or an embodiment combining both software and hardware aspects. Moreover, this application may take the form of a computer program product implemented on one or more computer-usable storage media that contain computer-usable program code, including but not limited to magnetic disk storage, CD-ROM, optical storage, etc. The schemes in the embodiments of this application can be implemented in various computer languages, such as object-oriented programming languages like Java and interpreted scripting languages like JavaScript.
This application is described with reference to the flowcharts and/or block diagrams of the methods, devices (systems), and computer program products in the embodiments of this application. It should be understood that each flow and/or block in the flowcharts and/or block diagrams can be implemented by computer program instructions. These computer program instructions can be provided to a processor of a general-purpose computer, special-purpose computer, embedded processor, or other programmable data processing device to create a machine, so that the instructions executed by the computer or other programmable data processing device processor generate devices for implementing the functions specified in one or more flows and/or blocks in the flowchart and/or block diagrams.
These computer program instructions can also be loaded onto a computer or other programmable data processing device, so that a series of operational steps are performed on the computer or other programmable device to provide steps for implementing the functions specified in one or more flows and/or blocks in the flowchart and/or block diagrams.
Described above are merely some preferred embodiments of this application, which are illustrative, and are not intended to limit this application. Though this application has been described in detail above, those skilled in the art can still make various modifications and replacements to the technical solutions recited in the embodiments provided herein. It should be understood that those modifications and replacements made without departing from the spirit of this application shall fall within the scope of this application defined by the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
202410006216.2 | Jan 2024 | CN | national |