The present disclosure relates to closed loop control of a position of a fifth wheel of a vehicle.
The commercial vehicle technology has grown drastically in the last five decades. Modern vehicles have superior control systems in the chassis in order to improve the performance and safety. Even though there are systems to control fifth wheel position, they are not intelligent enough to include advanced parameters. The fleet business and commercial vehicle transportation solutions have exponentially increased over the past few years. Therefore, performance criteria for fuel economy and traction are given more attention. The performance of the vehicle (Fuel economy, Stability and Tire wear) are highly dependent on aerodynamics and load on the drive wheels. For a specific application and a specific driving cycle (Acceleration, Maneuver style—rapid or smooth and Braking style) and specific environment condition (hill, road profile, wind speed and direction). The optimum fifth wheel position values to obtain necessary performance and meet legal regulations depend on the driving cycle (driver behavior) and vehicle application.
There are solutions to control the fifth wheel sliding for aerodynamics. But there is no solution for controlling it to include traction as well.
Therefore, there is a need for a method for closed loop control of a position of a fifth wheel of a vehicle which includes traction.
To that end, the present invention provides a method for determining a position of a fifth wheel on a vehicle, the method comprising the following steps:
According to these provisions, an appropriate fifth wheel position is determined.
According to an embodiment, the invention comprises one or more of the following features, alone or in any combination technically compatible.
According to an embodiment, the input features are collected by vehicle sensors placed on the vehicle and comprise at least one of vehicle driver information, vehicle state information, and environment information.
According to an embodiment, the vehicle driver information comprise at least one of throttle information, brake pedal information, steering angle information.
According to an embodiment, the vehicle state information comprise at least one of vehicle velocity, wheel speeds, axle load, GPS position of the vehicle, suspension articulation data, fuel consumed by the vehicle, acceleration and moments on the center of gravity of the vehicle, and a tire wear of at least one wheel of the vehicle, a torque applied to at least one wheel of the vehicle.
According to an embodiment, the environment information comprise at least one of hill angle, bank angle, wind speed, wind direction.
According to an embodiment, the input features further comprise at least one of slip data for the wheels of the vehicle, braking capacity, a road angle of ascent/descent for the road on which the vehicle is travelling, general engine data, road conditions such as wet, dry, icy, snowy, an acceleration/deceleration pattern over the period of time.
According to an embodiment, the method further comprises a step of providing the desired fifth wheel position as a target input for a closed loop control system to actuate fifth wheel actuators to move the fifth wheel according to the desired fifth wheel position.
According to an embodiment, the step of executing a second reinforcement learning model comprises a step of defining a vehicle parameter to be optimized by the desired fifth wheel position determined by the method.
According to an embodiment, the vehicle parameter is at least one of aerodynamics, i.e. fuel efficiency, traction, ride comfort, tire wear of the tires of the wheels, of the vehicle.
According to these provisions, the aerodynamics, the ride comfort, the traction and the tire wear are optimized on the vehicle through appropriate determination of the fifth wheel position.
According to an embodiment, inputs to the second reinforcement learning model comprise an estimated feedback on the vehicle parameter, the estimated feedback on the vehicle parameter indicating if the vehicle parameter was improved by a previous output of the second reinforcement learning model, the estimated feedback on the vehicle parameter being estimated by a vehicle parameter feedback estimation module based on vehicle information collected by a subset of the vehicle sensors placed on the vehicle to collect the input features.
According to these provisions, the second reinforcement learning model adapts and self-modifies over time, producing more accurate optimization of the vehicle parameter.
According to an embodiment, the method further comprises displaying a notification to manually modify the fifth wheel position of the vehicle based on the fifth wheel position.
According to another aspect, the invention provides a computer program apparatus, comprising a set of instructions configured to implement the method according to anyone of the embodiments described herein above, when the set of instructions are executed on a processor.
According to another aspect, the invention provides a vehicle comprising a fifth wheel and a processor and associated memory comprising the above computer program, the processor being configured to execute the computer program.
The foregoing and other purposes, features, aspects and advantages of the invention will become apparent from the following detailed description of embodiments, given by way of illustration and not limitation with reference to the accompanying drawings, in which the same reference refer to similar elements or to elements having similar functions, and in which:
Various embodiments of the disclosure are described in detail below. While specific implementations are described, it should be understood that this is done for illustration purposes only. Other components and configurations may be used without parting from the spirit and scope of the disclosure.
A fifth wheel is a hitch that allows to connect a cargo attachment to the back of a tracting vehicle, like a tractor or truck. For example, the fifth wheel refers to the “U” shaped coupling component found on the back of the towing vehicle, be it a large transport, pickup truck, or semi-truck.
Various examples and embodiments for defining an optimum position of a fifth wheel of a vehicle, based on a vehicle's drive cycle and the driver's driving style, are considered herein. Consider the following example: as a truck driver delivering ore from a mine drives from the mine to a foundry, the truck changes surfaces (dirt to asphalt to cement) and driving conditions (slow crawl with multiple turns at the mine, fast and straight on a highway, stop and go traffic before reaching the foundry). For optimal comfort, fuel efficiency, aerodynamics, tire wear, performance, or other desired vehicle applications, each of the conditions through which the truck will pass could have distinct optimal fifth wheel position FWP. In addition, how the driver operates the vehicle, and the driver's driving style or tendencies, could affect the optimal fifth wheel position FWP.
As illustrated in
The input features typically comprise vehicle driver information VDI, indicative of driver actions on the vehicle, vehicle state information VSI, indicative of vehicle dynamics, and environment information EI.
The vehicle driver information VDI comprise typically throttle pedal TH information, brake pedal BR information, steering wheel angle ST information.
The vehicle state information VSI comprise typically vehicle velocity V, wheel speeds, axle load, GPS position of the vehicle, suspension articulation data, fuel consumed by the vehicle, acceleration and moments on the center of gravity of the vehicle, and a tire wear of at least one wheel of the vehicle, a torque applied to at least one wheel of the vehicle.
The environment information EI comprise typically hill angle, bank angle, wind speed, wind direction, and GPS position of the vehicle.
Other exemplary data which can be collected could include slip data for the various wheels, braking capacity, angle of ascent/descent, general engine data, road conditions (wet, dry, icy, etc.), acceleration/deceleration patterns over a period of time, and/or any other data conveyed via the Controller Area Network (CAN) bus within the vehicle.
The outputs of the first reinforcement learning is provided to a second reinforcement learning model RL2, based on a second neural network, to determine the optimum position of a fifth wheel FWP, based on how the vehicle is currently being operated, i.e. based on the current driving cycle DC of the vehicle and the driving application DA for which the vehicle is currently being used, as determined by the first reinforcement learning model RL1. The second reinforcement learning model RL2 is also provided the input features IF collected by the vehicle sensors, so that the optimum position of a fifth wheel FWP is determined based on the input features IF collected by the vehicle sensors, and based on the current driving cycle DC of the vehicle and the driving application DA for which the vehicle is currently being used. Thereby, executing 103, via the processor aboard the vehicle, the second reinforcement learning model RL2, with inputs comprising the outputs of the first reinforcement learning model RL1 and the input features IF, outputs a desired fifth wheel position FWP.
In configurations where the vehicle is configured with actuators to self-modify the position of the fifth wheel during operation, the desired position of the fifth wheel is provided 104 as a target input for a closed loop control system CTL to actuate fifth wheel actuators FWA to move the fifth wheel according to the desired fifth wheel position FWP.
Any combination of the collected vehicle data can be input into a reinforcement learning model RL1, RL2 executed by a processor of the vehicle. The reinforcement learning model can also be a neural network, configured in a similar way as other neural networks.
One example output of the first reinforcement learning model RL1 can be a driving cycle of the vehicle, such as “transient” (where the vehicle is undergoing many changes, typical in stop and go traffic or off-roading), or “modal” (where the vehicle is going long periods of time at a constant speed). Another example output of the first reinforcement learning model RL1 can be an application of the vehicle, such as if the vehicle is being used to transport goods, ferry passengers, drive in an urban environment, drive off-road, etc.
Another input of the first reinforcement learning model RL1 is a feedback from the driver FFD of the vehicle. For example, as the vehicle sensors collect data and the reinforcement learning models RL1, RL2 analyze that data, the driver of the vehicle can confirm (or reject) the driving cycle DC and/or driving application DA of the vehicle. Over time, as the system collects additional data from the driver, the system can modify the first reinforcement learning model RL1 such that the driving cycle/vehicle application outputs are based on the feedback FFD provided by the vehicle driver. In this manner, the first reinforcement learning model RL1 adapts and self-modifies over time, producing more accurate predictions of the driving cycle and/or vehicle application. The system can also use GPS data, known road surface data, etc., to cross-check the feedback FFD received by the vehicle driver.
The outputs of the first reinforcement learning model RL1 are used as inputs to the second reinforcement learning model RL2, which is typically also based on a neural network. Additional inputs to the second reinforcement learning model RL2 also include the input features collected by vehicle sensors, which were used as inputs to the first reinforcement learning model RL1. In some configurations, the vehicle sensor data input to the second reinforcement learning model RL2 can be identical to the inputs of the first reinforcement learning model RL1, whereas in other configurations the vehicle sensor data input to the second reinforcement learning model RL2 can be a portion or subset of the vehicle sensor data input to the first reinforcement learning model RL1. In still other configurations, the vehicle sensor data input to the second reinforcement learning model RL2 can include vehicle data which was not used as an input to the first reinforcement learning model RL1. For instance, in some configurations, an additional input to the second reinforcement learning model RL2 can include a driver (or other human being) preference on how the vehicle should be optimized, i.e. on which vehicle parameter VP should be optimized, among for example aerodynamics, i.e. fuel efficiency, or traction, ride comfort, or tire wear of the tires of the wheels of the vehicle. In such cases, the driver can define 103b is that they desire that the fifth wheel of the vehicle be positioned to optimize the fuel economy of the vehicle. Other examples of driver preferences could be to minimize the wear on the tires of the wheels of the vehicle (thereby optimizing tire wear), optimizing ride comfort, or optimizing vehicle performance for a given scenario (such as optimizing for cornering versus optimizing for lack of cornering). While in this example the system can be configured to take a single driver preference, in other configurations the inputs to the second reinforcement learning model RL2 can also include multiple driver preferences, or ranked driver preferences; in particular, if the driver chooses «auto», the system would optimize the mentioned factors in a weighted way, such as 30% tires, 30% fuel economy etc. in a predefined scale. If the fuel level is low, the second reinforcement learning model RL2 would optimize mainly fuel economy. If the tire wear is more, it would increase the percentage for optimizing tire wear. If the DA is mining and load is more, it would optimize traction.
The second reinforcement learning model RL2 can then output optimal fifth wheel position values based on the inputs provided to the second reinforcement learning model RL2. In some circumstances, the second reinforcement learning model RL2 outputs values for all possible fifth wheel position values on a given vehicle each time the algorithm is executed, whereas in other circumstances the algorithm only outputs fifth wheel position values which vary from a current vehicle configuration.
Where the vehicle is equipped with actuators to adjust fifth wheel components, the fifth wheel position output of the second reinforcement learning model RL2 can be transmitted to one or more fifth wheel actuators FWA corresponding to the respective output, such that the fifth wheel actuators FWA adjust the vehicle fifth wheel position while the vehicle is in operation. If the actuator is broken or if the actuator is not able to drive the fifth wheel to the desired position a warning to the driver may be provided. In configurations where the vehicle is configured to auto-adjust while operating, the outputs of the second reinforcement learning model RL2 can be provided 104 to fifth wheel actuators FWA via control systems CTL within the vehicle. The actuators can then adjust the vehicle components according to the adjustment values output by the second reinforcement learning model RL2, and fifth wheel sensors FWS can compare the adjusted component values to the optimal/desired component values output by the second reinforcement learning model RL2. If adjusted component values detected do not match the desired values output by the second reinforcement learning model RL2, i.e. if an error ER is detected by the closed control loop, additional adjustments may be requested to the fifth wheel actuators FWA via the control system CTL within the vehicle.
Another input of the second reinforcement learning model RL2 is a feedback from the vehicle. For example, as the vehicle sensors collect data and the reinforcement learning models RL1, RL2 analyzes that data, some of or all of the vehicle data collected by the sensors are processed by a vehicle parameter feedback estimation VFE module to provide some estimated feedback FFV on at least one of the vehicle parameter VP being optimized by the second reinforcement learning model RL2. Over time, as the system provides additional feedback data FFV on the vehicle parameter VP being optimized, the system can modify the second reinforcement learning model RL2 such that the optimized vehicle parameter VP is based on the feedback FFV provided by the vehicle parameter feedback estimation VFE module. In this manner, the second reinforcement learning model RL2 adapts and self-modifies over time, producing more accurate optimization of the vehicle parameter VP.
For example, if traction is to be optimized, the system would use the following logic to check if traction has improved or not. Traction is determined as the force utilized to propel the vehicle. Consider the following newtons equation of motion
f=ma
Where f is the total force under the tires to propel the vehicle and a is the resulting acceleration in the center of gravity while m is the mass. The acceleration would be sensed using an inertial measurement unit in the vehicle and f would be estimated. The total force f can be increased if there is more load under the propelled tires. The fifth wheel location would be moved to provide optimum load on the rear wheel tires to generate increased total force f. For the same engine speed, same engine power, same tire and road conditions, increased total force f is generated due to more load without slipping the tires. This is optimized in low friction conditions like ice, wet etc. While, slow and go is a big driving application and especially for drivers with harsh acceleration and braking driving style.
Similarly, for fuel economy we will monitor the fuel consumed in periodic intervals for specific engine torque, throttle, road inclination etc.
Similarly, for tire wear we would rely on the tire wear information from existing tire sensors or slip information from the vehicle CAN bus using the wheel speed, vehicle speed and brake pedal or throttle pedal.
Similarly, for ride comfort we would monitor the accelerations on the cabin and seat preferably.
Hence, depending on the vehicle parameter VP which is optimized, a different subset of the vehicle sensors placed on the vehicle to collect the input features IF.
As an example of how to train the neural network which in turn is converted to executable code as a reinforcement learning model RL1, RL2, a vehicle manufacturer or other entity can collect known data, corresponding to the vehicle information and sensor data used as inputs to the reinforcement learning model RL1, RL2, such as: vehicle velocity, wheel speeds, steering angle, throttle, brake pedal depression, axle load data, GPS location, and/or suspension articulation data, hill angle, bank angle, wind speed, wind direction, etc. This feature data can be collected from multiple vehicles under multiple conditions, preferably with the amount of data collected from each vehicle being at least thirty minutes of operation, though the amount of data can vary.
In the case of the training of the first reinforcement learning model RL1, (1) the known input features data, (2) the corresponding, known driving cycles DC, (3) the corresponding, known driving applications DA, and (4) known fifth wheel positioning component values, can be compared via a sensitivity analysis, resulting in correlations between (1) the known feature data, (2) the corresponding, known driving cycles DC, (3) the corresponding, known driving applications DA. For example, the sensitivity analysis can execute models (such as a one-at a time test, a derivative-based local method, regression analysis, variance-based method, screening, scatter plots, etc.) to define how a given input/variable affects the likelihood of a specific condition (such as the X, Y, Z dimensions) in the fifth wheel position being determined. More specifically, the system can receive the known vehicle sensor data collected, driving cycles, vehicle applications, and determine how they affect the known fifth wheel position. The correlation outputs of the sensitivity analysis define the likelihood of a given variable affecting one or more of the outputs of the first reinforcement learning model RL1.
In the case of the training of the second reinforcement learning model RL2, (1) the known input features data, (2) the corresponding, known driving cycles DC, (3) the corresponding, known driving applications DA, and (4) known fifth wheel positioning component values, can be compared via a sensitivity analysis, resulting in correlations between (1) the known feature data, (2) the corresponding, known driving cycles DC, (3) the corresponding, known driving applications DA and (4) known fifth wheel positioning component values. For example, the sensitivity analysis can execute models (such as a one-at a time test, a derivative-based local method, regression analysis, variance-based method, screening, scatter plots, etc.) to define how a given input/variable affects the likelihood of a specific condition (such as the X, Y, Z dimensions) in the fifth wheel position being determined. More specifically, the system can receive the known vehicle sensor data collected, driving cycles, vehicle applications, and determine how they affect the known fifth wheel position. The correlation outputs of the sensitivity analysis define the likelihood of a given variable affecting one or more of the outputs of the second reinforcement learning model RL2, i.e. known fifth wheel positioning component values.
The outputs of the sensitivity analysis, as well the sensitivity analysis training data, can then be used to construct a machine learning network. For example, the correlations and test data associated with the sensitivity analysis can be input into Python, MatLab®, or other development software configured to construct neural network based on factor-specific data. Depending on the specific scenario, users can adjust the neural network construction by selecting from optimization methods including (but not limited to) the least-squares method, the Levenberg-Marquardt algorithm, the gradient descent method, or the Gauss-Newton method. The neural network can make predictions of the optimal fifth wheel position given input variables corresponding to the same data which were used to train the neural network. The neural network can then be converted to machine code and uploaded into memory, where upon execution by a processor the neural network operates as a machine learning model.
The training data can be driving data from various driving conditions, such as mining, long-haul, street, refuse, etc. The data from these driving conditions are evaluated via sensitivity analysis, and the resulting correlations can be used to construct the neural network for the first and second reinforcement learning model RL1, RL2. After initial constructions, the reinforcement learning model RL1, RL2 can operate based on a reward system. For example, considering the first reinforcement learning model RL1, after a day of driving, feedback from the driver can be collected. The system would ask the driver, via a driver feedback interface DFI module, a question regarding the driving conditions of the day, then use the answers the driver provides to determine if the model is accurately predicting driving conditions. For example, the system may ask the driver if they drove on rough surfaces with hard turns and an average velocity of 10 kilometers/hour, which is what the system predicted occurred. If the driver answers yes, the system gets a “reward,” meaning that the system further maintain the tune of the model in a similar manner going forward. If the driver answers no, the system will tune the model in a different direction until the driver begins answering “True” or “Yes” to the questions presented. More specifically, the Yes/No or True/False answers to the questions can modify the weights and biases of the connections between the nodes in the first reinforcement learning model RL1 neural network, where a yes/true can add weight to an existing connection, and a no/false can reduce weight of an existing connection. Questions can be presented to the drivers on a periodic basis (everyday, every week, etc.), when a change in conditions is detected (e.g., a change from dirt to asphalt), and/or when a certain amount of time driving has occurred (e.g. every four hours of driving).
Considering the second reinforcement learning model RL2, if the chosen option is to optimize fuel economy for a long-haul tractor-trailer in a flat terrain, and provided that the driving application and driving style is identified correctly by RL1, to be smooth accelerations and smooth braking which means the velocity change on the vehicle is not drastic and is a smooth profile, the fifth wheel position by RL2 would be set based on all the inputs, for example at 50 cm. The second reinforcement learning model RL2 would optimize the weights and bias based on the feedback. The feedback for fuel economy is checked in the vehicle. If the fuel economy has increased, for example in terms of miles per gallon, there would be a positive reward provided to RL2 and the weights and bias would modify towards that direction to get more accuracy, wherein weights and bias in the neural network of the second reinforcement learning model RL2 layer determine the output based on the inputs. If the fuel economy had decreased, then there would be a negative reward/punishment sent as a feedback to the second reinforcement learning model RL2 which would optimize the bias and weights in the opposite direction.
According to an aspect, the invention relates to a computer program apparatus, comprising a set of instructions configured to implement the method of anyone of the embodiments described herein above, when the set of instructions are executed on a processor.
According to a further aspect, the invention relates to a vehicle comprising a fifth wheel and a processor and associated memory comprising the above computer program, the processor being configured to execute the computer program.
Number | Date | Country | Kind |
---|---|---|---|
21209083.1 | Nov 2021 | WO | international |