Next Article in Journal
Design and Implementation of a Highly Efficient Quasi-Cyclic Low-Density Parity-Check Transceiving System Using an Overlapping Decoder
Previous Article in Journal
Time-Sensitive Networking to Improve the Performance of Distributed Functional Safety Systems Implemented over Wi-Fi
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Deep Reinforcement Learning Control Algorithm for Active Suspension Considering Uncertain Time Delay

1
School of Mechanical Engineering, Beijing Institute of Technology, Beijing 100811, China
2
State Key Laboratory of Automotive Simulation and Control, Jilin University, Changchun 130025, China
*
Author to whom correspondence should be addressed.
Submission received: 3 July 2023 / Revised: 31 August 2023 / Accepted: 7 September 2023 / Published: 12 September 2023
(This article belongs to the Section Physical Sensors)

Abstract

:
The uncertain delay characteristic of actuators is a critical factor that affects the control effectiveness of the active suspension system. Therefore, it is crucial to develop a control algorithm that takes into account this uncertain delay in order to ensure stable control performance. This study presents a novel active suspension control algorithm based on deep reinforcement learning (DRL) that specifically addresses the issue of uncertain delay. In this approach, a twin-delayed deep deterministic policy gradient (TD3) algorithm with system delay is employed to obtain the optimal control policy by iteratively solving the dynamic model of the active suspension system, considering the delay. Furthermore, three different operating conditions were designed for simulation to evaluate the control performance: deterministic delay, semi-regular delay, and uncertain delay. The experimental results demonstrate that the proposed algorithm achieves excellent control performance under various operating conditions. Compared to passive suspension, the optimization of body vertical acceleration is improved by more than 30%, and the proposed algorithm effectively mitigates body vibration in the low frequency range. It consistently maintains a more than 30% improvement in ride comfort optimization even under the most severe operating conditions and at different speeds, demonstrating the algorithm’s potential for practical application.

1. Introduction

With the development of microprocessor, sensor, and actuator technologies, scholars have studied various aspects of active suspension systems. The active suspension system can adaptively adjust the output force by controlling the actuator according to the driving operation and road conditions, achieving both ride comfort and better vehicle driving performance. The active suspension system control has significant prospects due to its ability to meet the comfort and safety requirements of modern vehicles.
Due to the closed-loop nature of the active suspension system, which includes sensors, controllers, and actuators, there are inherent delays in the measured signal or actuator. In the majority of cases, the time delay can be disregarded as it is insignificant. However, there are situations where the magnitude of the time delay is comparable to the control cycle, making it impossible to ignore and requiring careful consideration. Time delays tend to degrade control performance and potentially induce instability in the control system [1,2,3,4]. Therefore, for active suspension systems with delay, developing control strategies that resist uncertain time delays is particularly important.
Currently, delay-sensitive control systems have made significant progress in multiple industries including transportation [5,6], autonomous vehicle control [7], wireless sensor networks [8], and power systems [9,10]. In recent years, stability analysis and controller integration have been used in linear systems with measurement delays or actuator delays. In general, there are two main approaches to dealing with the problem of system delays. One is to design the controller using an integrated system model that includes actuator dynamics [11], and the other is to consider system delays during the controller design process. Ji et al. [12] proposed an improved variable universe fuzzy control strategy with real-time adjustment of the contracting–expanding factor parameters to improve the ride comfort of vehicles. Some scholars have proposed the use of delay compensation techniques to reduce or offset the negative effects of delays [13,14,15]. Udwadia et al. [16] proposed the application of delayed state positive feedback proportional control to the active control of structures. Pan et al. [17] designed a suspension delay active controller using an adaptive control strategy. Some scholars [18,19] have also attempted to design delayed instantaneous optimal control laws for suspension systems with delay using a state transformation approach to ensure the stability of the system. Du et al. [20] designed full-frequency domain state feedback controllers considering input delay for automotive suspensions and seat suspensions and achieved good damping effects within a certain delay range. Li et al. [21], on the other hand, proposed a time-delayed full-frequency domain robust control method based on dynamic output feedback. In addition, Kim et al. [22] combined perception technology and proposed a model predictive control of a semi-active suspension with a shift delay compensation using preview road information. Wu et al. [23] proposed a time-delay control strategy and the idea of using a linear motor as the actuator. According to this idea, they proposed a linear equivalent excitation method to optimize the optimal time-delay control parameters under complex excitation. Li et al. [24] proposed a fuzzy cooperative control strategy based on linear matrix inequality theory to weaken the effect of perturbations on vehicle driving conditions by integrating the problems of variable structure, disturbance immunity, and time delay. Based on Lyapunov theory and backstepping technique, Wang [25] studied the adaptive control problem of nonlinear active suspension systems with random perturbations and time delay. Moreover, some scholars have considered the time-delay characteristics of active suspension systems and controlled them according to the Takagi–Sugeno fuzzy model [26,27,28]. Further, scholars have been searching for more robust control methods to ensure the stability and performance of the suspension system [29,30,31].
Although numerous scholars have conducted extensive research on the problem of delay, the delay control system remains enigmatic due to its infinite-dimensional nature and the inherent uncertainty associated with delay. Therefore, further in-depth research is essential. Simultaneously, the automotive industry is rapidly adopting intelligence as part of its processes, which has led to the integration of artificial intelligence (AI) technology as a solution for controlling complex and uncertain systems. Notably, deep reinforcement learning (DRL) techniques have demonstrated significant advantages in addressing high-dimensional problems. By combining the robust perception and information-processing capabilities of deep learning (DL) with the decision-making proficiency of reinforcement learning (RL) in complex environments, these techniques are gradually gaining traction in the study of various intricate systems.
Currently, DRL has shown its advantages in vision and decision-making [32,33]. Some scholars have also tried to apply it to suspension control [34,35]. In an intuitive application, Pang et al. [36] proposed a non-fragile fault-tolerant control design for Markov-type systems. Providing more depth, Kozek et al. [37] proposed a neural algorithm based on reinforcement learning to optimize the creation of a linear quadratic regulator (LQR). In recent years, Li et al. [38] used an actor–critic architecture to study the adaptive neural network output feedback optimal control problem. In addition, many scholars have utilized different DRL architectures for active or semi-active suspension control. Lin et al. [39] studied a deep deterministic policy gradient (DDPG) control strategy for a full-vehicle active Macpherson suspension system. Yong et al. [40] proposed learning and control strategies for a semi-active suspension system in a full car using soft actor–critic (SAC) models on real roads. Similarly, Lee et al. [41] conducted a study on semi-active suspension control using DRL and proposed a state-normalization filter to improve the generalization performance. Further, Du et al. [42] proposed the utilization of external knowledge in the DDPG framework for suspension control and integration of speed planning to ensure ride comfort. It is worth mentioning that Han et al. [43] and Dridi et al. [44] have tried to use a proximal policy optimization (PPO) algorithm for semi-active and active suspension control and achieved satisfactory control results. Although there have been some useful studies on the application of DRL to suspension control, few attempts have been made to solve the delay problem. However, it is worth noting that in sequential decisions such as DRL control, the delay problem can have a significant impact on the actual control effect. Therefore, using DRL to solve the delay problem is still a new idea to be investigated. In this study, by adding a delay link to the twin-delayed deep deterministic policy gradient (TD3) algorithm, the agent is guided to explore the possibility of obtaining a more robust control strategy in a time-delayed environment, and then effectively suppress the effect of uncertain delay on the active suspension system. To summarize, the main innovations of this study are as follows:
  • To our knowledge, this study represents the first research endeavor to employ DRL techniques within the realm of delay control. The primary aim of this investigation is to alleviate the repercussions of uncertain delays through the implementation of active suspension control strategies rooted in DRL. Furthermore, this study demonstrates the utilization of high-dimensional advantages of DRL in an infinite-dimensional delay control system, ultimately achieving commendable results.
  • In this study, multiple sets of simulation considering deterministic delay, semi-regular delay, and uncertain delay are proposed to test the control performance of the algorithm. Various delay characteristics and uncertainty of the control system under their influence are considered to make the simulation closer to the actual working conditions.
  • The control algorithm proposed in this study maintains good control performance in multiple sets of the simulation built by MATLAB/Simulink for different working conditions and speeds, proving its application potential.
This paper is organized as follows. Section 2 presents the active suspension quarter model and road model. The proposed controller algorithm and the model design associated with it are presented in Section 3 and Section 4, respectively. In Section 5, the simulation environments with deterministic, semi-regular, and uncertain delays set up to validate the performance of the algorithm and their results are shown. In Section 6, the simulation results are discussed in a broader context with respect to the simulation results. Finally, Section 7 provides the conclusions of this study. A table of notations and abbreviations used in this paper is provided in Appendix A.

2. Dynamics Model of Active Suspension System Considering Time Delay

2.1. Active Suspension Quarter Model

In the field of control algorithms for active suspension systems, the two-degrees-of-freedom quarter suspension has emerged as the established benchmark model for experimentation. This model offers simplicity, and despite only addressing the vertical vibration of the sprung and unsprung masses, it provides an intuitive representation of the algorithm’s impact on control performance. Consequently, it aids researchers in algorithm development. Further, the vertical vibration of the vehicle body is the primary determinant of ride comfort, while the complex structure and pitch-and-roll motion of the body are presently disregarded. Hence, the quarter suspension serves as the foundational model for developing a dynamic model that accounts for delay.
Typically, models that take into account time delays are generally constructed from Takagi–Sugeno fuzzy models or include a model of an adjustable damper with hysteresis in the dynamics model [45,46,47]. The two-degrees-of-freedom active suspension dynamics model considering delay used in this study is shown in Figure 1. In this model, the system state delay is too small in magnitude, so it can be neglected, and only the inherent delay τ and actuator delay τ a in the control system are considered. Among them, the inherent delay is the amount of delay caused by the acquisition and transmission of signals in the active suspension control system and the controller operation, and the actuator delay is the amount of delay caused by the actuator response delay, which is unavoidable in the control loop.
According to the second class of Lagrangian equations, the kinetic equation with delayed quantities at moment t can be obtained as
{ m b x ¨ b ( t ) + k b ( x b ( t ) x u ( t ) ) + c b ( x ˙ b ( t ) x ˙ u ( t ) ) + u ( t τ τ a ) = 0 m u x ¨ u ( t ) + k b ( x u ( t ) x b ( t ) ) + k u ( x u ( t ) w ( t ) ) + c b ( x ˙ u ( t ) x ˙ b ( t ) ) u ( t τ τ a ) = 0
where m b is the sprung mass, kg. m u is the unsprung mass, kg. k b is the spring stiffness, N/m. k u is the equivalent tire stiffness, N/m. c b is the equivalent damping factor of the suspension damping element, N·s/m. x b is the vertical displacement of the sprung mass, m. x u is the vertical displacement of the unsprung mass, m. w is the vertical tire displacement, which can also be equated to the road displacement, m. F = u ( t τ τ a ) is the active suspension actuator control force for the time-delay system, N.
Take the system state variables as
x ( t ) = [ x b ( t ) x u ( t ) x ˙ b ( t ) x ˙ u ( t ) ] T
Take the system output variable as
y ( t ) = [ x ¨ b ( t ) x ˙ b ( t ) x ˙ b ( t ) x ˙ u ( t ) x b ( t ) x u ( t ) x u ( t ) w ( t ) ] T
Then, the state space expression of the active suspension, considering the delay, is
{ x ˙ ( t ) = A x ( t ) + B u ( t τ τ a ) + E w ( t ) y ( t ) = C x ( t ) + D u ( t τ τ a ) + L w ( t )
where A = [ 0 0 1 0 0 0 0 1 k b m b k b m b c b m b c b m b k b m u ( k b + k u ) m u c b m u c b m u ] , B = [ 0 0 1 m b 1 m u ] , E = [ 0 0 0 k u m u ] , C = [ k b m b k b m b c b m b c b m b 0 0 1 0 0 0 1 1 1 1 0 0 0 1 0 0 ] , D = [ 1 m b 0 0 0 0 ] , L = [ 0 0 0 0 1 ] .

2.2. Road Model

The vehicle road model uses a filtered white noise time domain road input model, i.e.,
w ˙ ( t ) = 2 π f 0 w ( t ) + 2 π S q ( n 0 ) v w 0 ( t )
where w ( t ) is the road displacement. f 0 is the lower cutoff frequency. S q ( n 0 ) is the road unevenness coefficient, which is related to the road class, where the unevenness coefficients of class A, B, and C roads are 16, 64, and 256, respectively. v is the speed, m/s. w 0 is a uniformly distributed white noise with a mean of 0 and an intensity of 1.
The lower cutoff frequency is calculated as
f 0 = 2 π n 00 v
where n 00 is the spatial cutoff frequency of the pavement, n 00 = 0.011   m 1 .
In this study, the time domain road unevenness curve used by the training agent is shown in Figure 2.

3. Controller Algorithm

This section introduces the primary algorithm employed in constructing the controller for the active suspension delay system. Initially, the fundamental principles and framework of reinforcement learning (RL) are presented, followed by an elucidation of the advantages offered by the deep reinforcement learning (DRL) algorithm compared to traditional RL algorithms. Subsequently, the TD3 algorithm, known for its suitability in continuous control systems, is selected based on the characteristics of the delay control system. The algorithmic process and technical intricacies of TD3 are then described.

3.1. Reinforcement Learning

RL is a payoff learning method developed from traditional attempted learning, and its basic principles can be traced back to optimal control theory and Markov decision processes (MDP). The RL can be represented as an interactive system consisting of an agent and an environment, as shown in Figure 3. At time t, the environment generates information describing the state of the system, which is the state s t . The agent interacts with the environment by observing the state and using this information to select the action a t . The environment accepts the action and transitions to the next state s t + 1 . The environment then feeds the next state and its reward r t to the agent. The cycle repeats and iterates continuously until the environment terminates. RL explores learning the optimal policy by maximizing the total reward in the process.

3.2. Deep Reinforcement Learning

In RL, the agent learns a function to formulate appropriate actions to maximize the goal. However, RL has limitations when dealing with large-scale problems. For example, in cases where the state space and action space are extremely large, traditional RL algorithms may face the issues of the curse of dimensionality and computational complexity. Additionally, traditional RL algorithms may require a large number of samples and time to learn a good policy, which can be time-consuming and inefficient for certain tasks. With the development of DL technology, we can use deep neural networks as function-approximation methods to learn value functions that imply multidimensional information. It extends the application of RL to higher dimensions, allowing it to suppress the high-dimensional disasters of traditional optimization problems to some extent.
The three primary functions that can be learned in RL align with the three primary methods in DRL, namely, policy-based, value-based, and model-based. Researchers use these three major classes of methods individually or in combination to meet the needs of practical tasks. In this study, the active suspension control system considering delay is a typical continuous state and continuous action control system. The classical DRL algorithm in this field is the DDPG [48]. DDPG utilizes an actor–critic architecture for learning, i.e., an actor network is added to the deep Q network (DQN) [49,50] to output action values directly. It helps the agent to obtain more feedback information and thus make more accurate decisions.

3.3. Twin-Delayed Deep Deterministic Policy Gradient

Although the classical DDPG algorithm has achieved satisfactory results in many tasks, the traditional actor–critic framework suffers from the bias and variance problems associated with function approximation, considering the cumulative nature of the DRL iterative process. Specifically, on the one hand, the variance can cause overestimation, while at the same time, high variance can cause the accumulation of errors, which in turn makes the system less stable. Considering the characteristics of time-delayed control systems, high variance means more aggressive control behavior and high deviation means gradual deviation from the steady state, and the agent gains more benefit but also takes more risk, which is more likely to cause instability of the control system. Therefore, we need to suppress both overestimation and cumulative error. Based on the above practical control requirements, this study uses the TD3 algorithm [51] as the basic algorithmic framework, i.e., adding techniques such as clipped double Q-learning, delayed policy updates, and target policy smoothing to the DDPG framework. The framework improvement of the TD3 algorithm is shown in Figure 4.
When building a controller using the TD3 algorithm, first, randomly initialize the critic network Q 1 ( s , a | θ Q 1 ) , Q 2 ( s , a | θ Q 2 ) , and actor network μ ( s | θ μ ) with parameters θ Q 1 , θ Q 2 , and θ μ , respectively. At the same time, we need to initialize target networks Q 1 ( s , a | θ Q 1 ) , Q 2 ( s , a | θ Q 2 ) , and μ ( s | θ μ ) with parameters θ Q 1 , θ Q 2 , and θ μ , respectively. Then, synchronize the parameters:
θ Q 1 θ Q 1 , θ Q 2 θ Q 2 , θ μ θ μ
In the learning phase, we set the capacity of the experience buffer and the number of episodes. After initializing the state s t , the actions are selected based on the current actor network and exploration noise:
a t = μ ( s t | θ μ ) + ξ t
where ξ t N t ( 0 , σ ) .
Performing actions in the environment and obtaining rewards r t and new state s t + 1 . We store these experiences in the form of samples of ( s t , a t , r t , s t + 1 ) that are transformed into the experience buffer. When the experience buffer reaches a certain number of samples, we randomly sample a mini-batch of experience–transfer samples ( s i , a i , r i , s i + 1 ) from it to perform parameter updates. Let
y i = r i + γ min j = 1 , 2 Q j ( s i + 1 , μ ( s i + 1 | θ μ ) | θ Q j + ζ ) , ζ c l i p ( N ( 0 , σ ) , c , c )
where γ is the discount factor and takes the value of 0.99, which is taken to be equivalent to considering the situation after 100 time steps. c is the noise clipping, which in this study is taken as 0.5. The clipped double Q-learning idea and target policy smoothing idea of TD3 are reflected here. With clipped double Q-learning, the value target cannot introduce any additional overestimation bias using the standard Q-learning target. While this update rule may induce an underestimation bias, this is far preferable to overestimation bias, as unlike overestimated actions, the value of underestimated actions will not be explicitly propagated through the policy update [51]. Moreover, the effect of inaccuracy caused by function approximation error can be effectively reduced by target policy smoothing.
Update the parameters θ Q j of the critic network according to the minimization loss function, i.e.,
θ Q j L ( θ Q j ) = 1 N i ( y i Q j ( s i , a i | θ Q j ) ) 2 θ Q j Q j ( s i , a i | θ Q j ) , j = 1 , 2
where N is the size of the mini-batch sample, whose value we take as 128 in this study. According to our experience, the size of the mini-batch should be related to the complexity of the problem being studied. The parameters θ μ of the actor network are updated according to the objective maximization function, i.e.,
θ μ J ( μ ) = 1 N i ( a Q ( s , a | θ Q ) | s = s i , a = μ ( s i ) θ μ μ ( s | θ μ ) | s i )
Finally, in order to avoid unknown oscillations affecting convergence during gradient descent to ensure that the update of the network can balance stability and rapidity, we perform the soft update process for the network parameters, i.e.,
θ Q j ε θ Q j + ( 1 ε ) θ Q j , j = 1 , 2 θ μ ε θ μ + ( 1 ε ) θ μ
where ε is the soft update factor, since the delay control system is a high-dimensional unstable system, ε = 0.001 is taken in this study to satisfy the robustness of the policy to some extent.
In addition, the critic network is updated twice as often as the actor network in order to first minimize the errors induced by the value estimation before introducing policy updates. Soft updating of network parameters and delayed policy updating ensure target stabilization and thus reduce error increase. TD3 is summarized in Algorithm 1.
Algorithm 1 TD3
1Randomly initialize critic networks Q 1 ( s , a | θ Q 1 ) , Q 2 ( s , a | θ Q 2 ) , and actor network μ ( s | θ μ ) with random parameters θ Q 1 , θ Q 2 , and θ μ .
2Initialize target networks θ Q 1 θ Q 1 , θ Q 2 θ Q 2 , θ μ θ μ
3Initialize replay buffer Ω
4for episode = 1 to Max episodes, do
5 Initialize a Gaussian random process ξ t N t ( 0 , σ ) for action exploration.
6 Receive the initial observation state s 0 of the environment.
7for t = 1 to Max steps, do
8  Select action a t = μ ( s t | θ μ ) + ξ t according to the current policy and exploration noise.
9  Execute action a t and observe reward r t and observe the new state s t + 1 .
10  Store transition tuple ( s t , a t , r t , s t + 1 ) in Ω
11  Sample mini-batch of N transitions ( s i , a i , r i , s i + 1 ) from Ω
12  Set y i = r i + γ min j = 1 , 2 Q j ( s i + 1 , μ ( s i + 1 | θ μ ) | θ Q j + ζ ) , ζ c l i p ( N ( 0 , σ ) , c , c )
13  Update parameters θ Q j : θ Q j L ( θ Q j ) = 1 N i ( y i Q j ( s i , a i | θ Q j ) ) 2 θ Q j Q j ( s i , a i | θ Q j ) , j = 1 , 2
14  if t mod delayed update frequency, then
15   Update the parameters θ μ : θ μ J ( μ ) = 1 N i ( a Q ( s , a | θ Q ) | s = s i , a = μ ( s i ) θ μ μ ( s | θ μ ) | s i )
16   Soft update target networks: θ Q j ε θ Q j + ( 1 ε ) θ Q j , j = 1 , 2 θ μ ε θ μ + ( 1 ε ) θ μ
17  end if
18end for
19end for

4. Controller Model

Solving a new problem using DRL involves the creation of an environment, so in this section, we focus on the environment components involved in the controller model, namely, the state, action, reward, and transfer function. Among them, the transfer function of DRL is the dynamics model introduced in Section 2, which will not be repeated here.

4.1. State

The state is the information that describes the environment, and the RL environment must provide the algorithm with enough information to solve the problem. In the active suspension delay control system, multidimensional information such as displacement, velocity, acceleration, control force, and time are included. These high-level states are the raw states. The raw state should contain all the information relevant to the problem, but it is often difficult to learn because it usually has a lot of redundancy and background noise. Raw and complete information provides more freedom, but extracting and interpreting useful signals from it is a much heavier burden. Therefore, the selection and design of the state space becomes particularly important.
Based on the important suspension performance parameters included in the dynamics model, the following state variables are chosen to characterize the state space considering the control performance requirements of the active suspension. The designed state space takes into account both real-world constraints and observability cases under the influence of sensor arrangements.
s = [ x ¨ b x ˙ b x ˙ b x ˙ u x b x u ] T
The state information takes into account the ride comfort, actuator efficiency, and suspension travel of the suspension system. This state contains much less information than the original state, but it contains more straightforward information for the algorithm and is easier to learn.
Further, the agent further preprocesses the designed states for its own use. In order to improve the generalization performance of the controller model, this study normalizes the states, and the final state vector is represented as
s = [ x ¨ b λ 1 x ˙ b λ 2 x ˙ b x ˙ u λ 3 x b x u λ 4 ] T
where λ is the normalized coefficient of each state variable.

4.2. Action

Action is the output of the agent, which is the amount of control output by the controller. Actions change the environment by transitioning the dynamics model to the next state and into the next round of iterations. How the action is designed affects the ease of control of the system and thus directly affects the problem’s difficulty. In this study, the action is the control force of the active suspension actuator, i.e.,
F t = a t = μ ( s t | θ μ )
where μ ( s | θ μ ) is the actor network. It should be noted that due to the delay, the control force in the actual system often appears as
a t + τ + τ a = μ ( s t | θ μ )
Considering the specific performance constraints of the actuator, adding the double truncation constraint to the actor network, the final action is represented as
a = c l i p ( a t + τ + τ a , F min , F max )
In this study, based on the saturation constraint of the actuator, we set F min = 3   kN and F max = 3   kN .

4.3. Reward

The reward signal is used to define the objective that the agent should maximize to guide the agent’s exploration to obtain a better policy from a global level. The reward function can be somewhat analogous to the objective function of a traditional control problem, so reward design is an important issue in DRL. In the design of an active suspension controller considering delay, the control performance requirements of multiple objectives need to be considered.
The first issue is the ride comfort, which is closely related to the vertical acceleration of the vehicle body. Therefore, the impact of road unevenness on the body is avoided, and the vertical acceleration is reduced while a suitable active control force is needed.
Secondly, the practical constraints of suspension travel need to be considered. Dynamic suspension travel needs to satisfy the following inequalities within a safe range:
| x b ( t ) x u ( t ) | f d
where f d is the maximum travel value of the suspension, and in this study, let f d = 0.15   m . The value is determined by referring to the literature [52] and considering the actual constraints of the suspension obtained.
Then, the grounding of the suspension needs to be considered, i.e., the following inequalities need to be satisfied to ensure the handling stability of the vehicle:
k u | x u ( t ) w ( t ) | F m
where F m is the static load of the tire, and its calculation formula is
F m = ( m b + m u ) g
Finally, the control characteristics of the actuator need to be considered. The actuator delay τ a has a close relationship with the total delay of the whole system, so in order to suppress the effect of delay at the physical level, we should ensure that the control force is relatively stable in a relatively small interval as much as possible.
In summary, the reward function is defined as
r = ( k 1 | x ¨ b | 2 + k 2 | x b x u | 2 + k 3 | x u w | 2 + k 4 | F | 2 )
where k 1 = 0.7 , k 2 = 0.1 , k 3 = 0.1 , k 4 = 0.1 are the weight coefficients of the balanced multi-objective optimization problem.
It should be noted that the agent’s reward function in the training phase references the state information of the system after the control force has been applied after the delay. In contrast, the state referenced by the actor network and critic network is the current state. In other words, the delayed control system in the experimental phase is equivalent to an open-loop control system.

5. Simulation and Results

This section describes the simulation and results used to verify the proposed control algorithm in the simulation environment built by MATLAB2023a/Simulink. First, the specific environment and agent-related network information required for the simulation experiments were introduced and set up. Then, parallel experiments were conducted for 10 ms, 20 ms, and 30 ms defined delay conditions to demonstrate the control performance of the proposed algorithm for deterministic delay control. In addition, we established a semi-regular delay control condition based on the fuzzy relationship between actuator control force and delay to test the excellent control performance of the proposed algorithm. Finally, we established a severe operating condition with uncertain delay to test the anti-disturbance performance of the proposed algorithm and its improvement of ride comfort.

5.1. Implementation Details

The active suspension system dynamics model with consideration of delay introduced in Section 2 was used to build the environment required for training the DRL, and the model parameters and values are listed in Table 1. The road information used for training the agent is shown in Figure 2, and a deterministic delay of 30 ms was added to the network during training.
The critic and actor network used for the agent were specifically designed for the control of active suspension systems, as shown in Figure 5, and the hyperparameters used to train the network are shown in Table 2. In order to better verify the performance of the proposed algorithm in this study, we chose the active suspension DDPG control architecture proposed in the literature [53] as a baseline for comparison. The hyperparameters of the baseline algorithm were used in this study, and additional hyperparameters were selected based on the original TD3 algorithm. In addition, we performed combinatorial experiments on some of the hyperparameters; see Appendix B.

5.2. Deterministic Delayed Conditions

Due to the random nature of the road excitation during the actual driving of the vehicle, we selected a random class C road as the random excitation. The proposed active control algorithm of the delay system was used to study the dynamics of the suspension system under random excitation. One study [20] showed that the classical control method could not prevent unstable divergence in an active suspension system with a 24 ms time delay. Based on this conclusion, it was determined that the delay times were designed to be similar to 10 ms, 20 ms, and 30 ms.
We quantitatively evaluated the intensity of body vibration under DRL control using root mean square (RMS) values of acceleration. In addition, we also analyzed the frequency response of the body acceleration to a random road. The RMS results of the proposed algorithm with different time delays in the simulation environment were compared with passive results, as tabulated in Table 3. The simulation comparison results of the body acceleration and the frequency response are shown in Figure 6. In addition, we compared the proposed algorithm with the most classical and effective DDPG algorithm [53]; the comparison results are also presented in the graphs.
As we can see from the graphs, the proposed control algorithm optimized the ride comfort by 43.58%, 44.9%, and 32.28% for 10 ms, 20 ms, and 30 ms deterministic delays, respectively, compared to the passive suspension. Although the control algorithm of the proposed algorithm is slightly inferior to that of DDPG under the low latency condition of 10 ms, the control performance of the proposed algorithm improves by 25.56% compared to that of DDPG under the latency condition of 20 ms. Further, under the large delay condition of 30 ms, the proposed algorithm still maintains the optimization result of 32.28% compared to DDPG, which cannot maintain stability and crashes. The above results clearly demonstrate the superior control performance of the proposed algorithm. Although the proposed algorithm exhibited good control performance at deterministic delays, realistic control systems did not always have deterministic delays. Deterministic delay is equivalent to adding a deterministic dimension to the overall control system, which is still solvable to some extent. Therefore, the study of delayed control systems requires more in-depth discussion and analysis.

5.3. Semi-Regular Delayed Conditions

The amount of delay in a control system is often closely related to the actual actuation capability of the actuator, so some scholars used an integrated system model that includes the actuator dynamics to design the controller. To simulate this characteristic, we developed a semi-regular delay condition, specifically using the following rules:
τ a = { δ , 0 < | Δ F | f 2 δ , f < | Δ F | 2 f 3 δ , 2 f < | Δ F | 3 f
where δ is the unit delay amount, and its value was taken as 10 ms in this study. f is the unit amount determined according to the maximum limiting control force of the actuator, and its value was taken as 2 kN in this study. This value is chosen by considering the bandwidth of the active suspension actuator and can actually be obtained by testing the actuation force response characteristics of the active suspension actuator under different loads.
The semi-regular delay condition was based on the fuzzy relationship between actuator delay and actuation capacity, and the delay was graded in steps. It can make the system simple while retaining the actuator’s role to a greater extent, so this condition had stronger practical significance for testing the control algorithm. The body acceleration and frequency response with the semi-regular delay condition are shown in Figure 7.
In the semi-regular delay condition, the RMS result of the proposed algorithm is 0.9529 m/s2, and the ride comfort is optimized by 44.13%. In comparison, the RMS value under the DDPG baseline control is 1.1321 m/s2, while the control performance of the proposed algorithm exceeds the DDPG baseline by 15.8%. We can see that the proposed algorithm still maintained good control performance in the operating conditions where the fuzzy characteristics of the actuator were considered.

5.4. Uncertain Delay Conditions

The previous consideration of deterministic delay and semi-regular delay conditions amounted to adding one or more dimensions of uncertainty to the delay control system. Further, the delay in a practical time-delay control system was often uncertain. Mathematically and theoretically, this environment was closer to an infinite-dimensional control system. Therefore, we set up simulation experiments under uncertain delay time working conditions to simulate a more severe environment. Under uncertain delay conditions, DDPG fails to converge due to its inability to adapt to such a high-dimensional working condition.
Specifically, the delay time was set to satisfy the uniform distribution of the following equation:
τ a = ν , ν U ( 10 , 40 )
The body acceleration and frequency response under the uncertain delay conditions are shown in Figure 8. The value of RMS under the control of the proposed algorithm was 1.1116 m/s2 in the uncertain delay condition, and the optimization of ride comfort reached 37.56%. In comparison, the RMS value under the DDPG baseline control is 1.2528 m/s2, while the control performance of the proposed algorithm exceeds the DDPG baseline by 11.3%. It can be seen that the control performance of the proposed algorithm only showed a small degradation in the case of a sharp increase in system complexity. In Figure 8b, the response of the proposed algorithm to road vibration shows a slight increase in a certain range in the high-frequency part. This is due to the fact that the uncertain delay condition is an extremely severe condition, so the DRL algorithm will prioritize the control needs in the low-frequency section. A more in-depth study will be carried out for high-frequency control.
It should be noted that in the comparative experiments, the DDPG algorithm was unable to complete the full verification due to triggering the termination condition within the episode described in Equation (18) under the conditions of a 30 ms large delay, a semi-regular delay, and an uncertain delay. In order to conduct the comparative experiments, we had to remove the relevant termination conditions. In comparison, the proposed algorithm consistently did not trigger the termination condition, ensuring the safety of the driving process. Furthermore, in order to verify the generalization performance of the proposed algorithm in different environments, we conducted experiments by varying the speed. The control results are shown in Table 4. The table indicates that the proposed algorithm shows good control performance at different speeds.

6. Discussion

In this study, we made some beneficial attempts using DRL to solve the challenging problem of time delay control in active suspension systems. We set deterministic, semi-regular, and uncertain time delays to simulate the changes from ideal working conditions to real working conditions and then to harsh working conditions, thereby testing the control performance of the proposed algorithm. Under deterministic delay, the proposed algorithm demonstrated good control performance at working conditions of 10 ms, 20 ms, and 30 ms, surpassing the DDPG baseline and maintaining good stability even under larger time delays. In addition, the proposed algorithm effectively suppressed the first resonance peak of road excitation and body, and improved ride comfort. The proposed algorithm includes predictions of future rewards, thus possessing stronger robustness to a certain extent. This condition corresponds to a relatively ideal working environment for the actuator, where stable fixed delay is desirable. However, under actual conditions, system delay is often closely related to the actuator’s manufacturing capability and response. Therefore, we designed semi-regular delay conditions to simply simulate this characteristic. The simulation results also reflected the good control performance of the proposed algorithm and its improvement in ride comfort. We believe that these results are due to the fact that the proposed algorithm, based on predicting the future, has imposed planning on the output of control force, keeping it within a small range of variation that better aligns with the actuator’s response characteristics. Furthermore, uncertain conditions are relatively harsh working conditions, and it is necessary to conduct simulations under such conditions to better test the performance of the proposed algorithm. It can be seen that under such conditions, the proposed algorithm can still maintain a 37.56% optimization effect. We believe this is because, for the infinite-dimensional delay control system, the data-driven algorithm bypasses the analysis at the high-dimensional level and directly performs end-to-end analysis. To a certain extent, it corresponds to re-architecting a solver that simulates a high-dimensional environment, and this approach is undoubtedly novel and effective. Of course, further research is needed to verify its effectiveness. It is encouraging that Baek et al. [54] and Li et al. [55] have attempted to apply DRL to robotics, Zhu et al. [56] have applied it in Mobile Sensor Networks, and Chen et al. [57] have generalized it even more to continuous control. At the same time, it is also important for us to focus on choosing more suitable algorithms for the control of delay systems. In recent years, PPO [58] has been widely applied in the industry due to its robustness and performance. The characteristics of PPO in parameter tuning and policy updating provide us with new ideas for our future research, which will be a valuable direction for future studies.
Furthermore, the generalization ability of the algorithm has always been a controversial issue for learning-based algorithms. For this reason, we conducted simulations at different speeds, and the results showed that the proposed algorithm maintained over 30% comfort optimization from 10 m/s to 50 m/s, which covers almost all driving speeds in reality. Moreover, to apply the proposed algorithm in the real world, the complexity of the algorithm must be considered, and its real-time calculation performance must be examined. The trained DRL controller has 33,537 parameters, and it only takes 5.6 ms to compute on a typical PC (Intel Core i9-12900KF, 16 GB RAM). Therefore, the proposed controller will not be a problem in terms of real-time implementation.

7. Conclusions

This paper proposed an active suspension DRL control algorithm considering time delay to study the uncertain delay problem in the actual control system of active suspension. Firstly, a dynamics model of the active suspension system considering time delay was established. Secondly, the TD3 algorithm was enhanced by incorporating delay, enabling the agent to explore more robust policies. Finally, simulation experiments were conducted under three different experimental conditions: deterministic delay, semi-regular delay, and uncertain delay. The proposed algorithm’s control performance was evaluated, and experimental validation was performed at various speeds. The results illustrate the algorithm’s effectiveness in mitigating the impact of uncertain delay on the active suspension system, resulting in significant improvements in ride comfort optimization. Specifically, the proposed algorithm achieved comfort optimization rates of 43.58%, 44.9%, and 32.28% for deterministic delays of 10 ms, 20 ms, and 30 ms, respectively. Additionally, it obtained optimization rates of 44.13% and 37.56% for semi-regular and uncertain delay conditions, respectively. Furthermore, when compared to the DDPG baseline algorithm, the proposed algorithm demonstrates excellent stability and convergence even under complex delay conditions.
Despite satisfactory results in the current research, the important characteristic of time delay still requires further investigation. In future work, we aim to enhance our understanding of the relationship between delay and control performance by incorporating an integrated system model that accounts for actuator dynamics into the DRL environment. By relying on a comprehensive model environment, the agent can derive improved control policies that are better suited for real-world vehicle deployment scenarios.

Author Contributions

Conceptualization, Y.W., C.W., S.Z. and K.G.; methodology, Y.W. and C.W.; software, Y.W.; validation, Y.W. and S.Z.; formal analysis, Y.W.; investigation, Y.W.; resources, Y.W.; data curation, Y.W.; writing—original draft preparation, C.W.; writing—review and editing, Y.W. and S.Z.; visualization, Y.W.; supervision, K.G.; project administration, K.G.; funding acquisition, K.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The National Key Research and Development Program of China (Grant No. 2022YFB3206602) and China Postdoctoral Science Foundation Funded Project (Grant No. 2022M720433).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors sincerely thank the anonymous reviewers for their critical comments and suggestions for improving the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Notations and Abbreviations

Table A1. The notations used in the paper.
Table A1. The notations used in the paper.
ParametersNotationUnit
τ Inherent delay s
τ a Actuator delay s
m b Sprung mass kg
m u Unsprung mass kg
k b Spring stiffness N / m
k u Equivalent tire stiffness N / m
c b Equivalent damping factor N s / m
x b Vertical displacement of the sprung mass m
x u Vertical displacement of the unsprung mass m
w Road displacement m
F = u ( t τ τ a ) Active suspension actuator control force N
v Speed m / s
f 0 Lower cutoff frequency
S q ( n 0 ) Road unevenness coefficient
w 0 Uniformly distributed white noise
n 00 Spatial cutoff frequency
s t State at time t
a t Action at time t
s t + 1 State at time t + 1
r t Reward at time t
Q ( s , a | θ Q ) Critic network
μ ( s | θ μ ) Actor network
Q ( s , a | θ Q ) Target critic network
μ ( s | θ μ ) Target actor network
γ Discount factor
ε Soft update factor
ξ t Exploration noise
ζ Smoothing noise
λ Normalized coefficient
f d Maximum travel value of the suspension
F m Static load of the tire
k Weight coefficient
δ Unit delay amount s
f Unit amount determined force kN
Table A2. The list of abbreviations.
Table A2. The list of abbreviations.
AbbreviationFull Name
DDPGDeep Deterministic Policy Gradient
DLDeep Learning
DQNDeep Q Network
DRLDeep Reinforcement Learning
LQRLinear Quadratic Regulator
MDPMarkov Decision Processes
PPOProximal Policy Optimization
RLReinforcement Learning
RMSRoot Mean Square
SACSoft Actor–Critic
TD3Twin-Delayed Deep Deterministic Policy Gradient

Appendix B. Hyperparametric Portfolio Experiment

In this study, a specifically designed deep neural network architecture was chosen to reduce computational complexity to meet the requirements of the controller in the real world. In order to address the exploration and utilization challenges, the learning rate of actor and critic networks is also an important influence. Therefore, we have investigated the hyperparameter combinations of these two components separately, which are shown in Table A3 and Table A4, respectively.
Table A3. Number of neurons in the deep neural network.
Table A3. Number of neurons in the deep neural network.
Neuronal Assemblies32–6464–128128–256256–512
Optimization24.99%30.30%35.62%32.82%
Table A4. Critic network and actor network learning rate. The first behavioral critic network learning rate and the first column is the actor network learning rate.
Table A4. Critic network and actor network learning rate. The first behavioral critic network learning rate and the first column is the actor network learning rate.
Optimization0.10.010.001
0.116.77%23.79%23.72%
0.0128.59%31.98%28.82%
0.00131.00%35.62%32.81%

References

  1. Yan, G.H.; Wang, S.H.; Guan, Z.W.; Liu, C.F. PID Control Strategy of Vehicle Active Suspension Based on Considering Time-Delay and Stability. Adv. Mater. Res. 2013, 706–708, 901–906. [Google Scholar] [CrossRef]
  2. Xu, J.; Chung, K.W. Effects of Time Delayed Position Feedback on a van Der Pol–Duffing Oscillator. Phys. D Nonlinear Phenom. 2003, 180, 17–39. [Google Scholar] [CrossRef]
  3. Zhang, H.; Wang, X.-Y.; Lin, X.-H. Topology Identification and Module–Phase Synchronization of Neural Network with Time Delay. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 885–892. [Google Scholar] [CrossRef]
  4. Min, H.; Lu, J.; Xu, S.; Duan, N.; Chen, W. Neural Network-Based Output-Feedback Control for Stochastic High-Order Non-Linear Time-Delay Systems with Application to Robot System. IET Control. Theory Appl. 2017, 11, 1578–1588. [Google Scholar] [CrossRef]
  5. Chen, X.; Leng, S.; He, J.; Zhou, L.; Liu, H. The Upper Bounds of Cellular Vehicle-to-Vehicle Communication Latency for Platoon-Based Autonomous Driving. IEEE Trans. Intell. Transp. Syst. 2023, 24, 6874–6887. [Google Scholar] [CrossRef]
  6. Li, J.; Liu, X.; Xiao, M.; Lu, G. A Planning Control Strategy Based on Dynamic Safer Buffer to Avoid Traffic Collisions in an Emergency for CAVs at Nonsignalized Intersections. J. Transp. Eng. Part A Syst. 2023, 149, 04023066. [Google Scholar] [CrossRef]
  7. Xu, L.; Ma, J.; Zhang, S.; Wang, Y. Car Following Models for Alleviating the Degeneration of CACC Function of CAVs in Weak Platoon Intensity. Transp. Lett. 2023, 15, 1–13. [Google Scholar] [CrossRef]
  8. Samiayya, D.; Radhika, S.; Chandrasekar, A. An Optimal Model for Enhancing Network Lifetime and Cluster Head Selection Using Hybrid Snake Whale Optimization. Peer-to-Peer Netw. Appl. 2023, 16, 1959–1974. [Google Scholar] [CrossRef]
  9. Reddy, P.Y.; Saikia, L.C. Hybrid AC/DC Control Techniques with Improved Harmonic Conditions Using DBN Based Fuzzy Controller and Compensator Modules. Syst. Sci. Control Eng. 2023, 11, 2188406. [Google Scholar] [CrossRef]
  10. Wang, R.; Jorgensen, A.B.; Liu, W.; Zhao, H.; Yan, Z.; Munk-Nielsen, S. Voltage Balancing of Series-Connected SiC Mosfets with Adaptive-Impedance Self-Powered Gate Drivers. IEEE Trans. Ind. Electron. 2023, 70, 11401–11411. [Google Scholar] [CrossRef]
  11. Klockiewicz, Z.; Slaski, G. Comparison of Vehicle Suspension Dynamic Responses for Simplified and Advanced Adjustable Damper Models with Friction, Hysteresis and Actuation Delay for Different Comfort-Oriented Control Strategies. Acta Mech. Autom. 2023, 17, 1–15. [Google Scholar] [CrossRef]
  12. Ji, G.; Li, S.; Feng, G.; Wang, H. Enhanced Variable Universe Fuzzy Control of Vehicle Active Suspension Based on Adaptive Contracting-Expanding Factors. Int. J. Fuzzy Syst. 2023, 1–15. [Google Scholar] [CrossRef]
  13. Han, S.-Y.; Zhang, C.-H.; Tang, G.-Y. Approximation Optimal Vibration for Networked Nonlinear Vehicle Active Suspension with Actuator Time Delay. Asian J. Control. 2017, 19, 983–995. [Google Scholar] [CrossRef]
  14. Lei, J. Optimal Vibration Control of Nonlinear Systems with Multiple Time-Delays: An Application to Vehicle Suspension. Integr. Ferroelectr. 2016, 170, 10–32. [Google Scholar] [CrossRef]
  15. Bououden, S.; Chadli, M.; Zhang, L.; Yang, T. Constrained Model Predictive Control for Time-Varying Delay Systems: Application to an Active Car Suspension. Int. J. Control Autom. Syst. 2016, 14, 51–58. [Google Scholar] [CrossRef]
  16. Udwadia, F.E.; Phohomsiri, P. Active Control of Structures Using Time Delayed Positive Feedback Proportional Control Designs. Struct. Control. Health Monit. 2006, 13, 536–552. [Google Scholar] [CrossRef]
  17. Pan, H.; Sun, W.; Gao, H.; Yu, J. Finite-Time Stabilization for Vehicle Active Suspension Systems with Hard Constraints. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2663–2672. [Google Scholar] [CrossRef]
  18. Yang, J.N.; Li, Z.; Danielians, A.; Liu, S.C. Aseismic Hybrid Control of Nonlinear and Hysteretic Structures I. J. Eng. Mech. 1992, 118, 1423–1440. [Google Scholar] [CrossRef]
  19. Kwon, W.; Pearson, A. Feedback Stabilization of Linear Systems with Delayed Control. IEEE Trans. Autom. Control. 1980, 25, 266–269. [Google Scholar] [CrossRef]
  20. Du, H.; Zhang, N. H∞ Control of Active Vehicle Suspensions with Actuator Time Delay. J. Sound Vib. 2007, 301, 236–252. [Google Scholar] [CrossRef]
  21. Li, H.; Jing, X.; Karimi, H.R. Output-Feedback-Based Hınfty Control for Vehicle Suspension Systems with Control Delay. IEEE Trans. Ind. Electron. 2014, 61, 436–446. [Google Scholar] [CrossRef]
  22. Kim, J.; Lee, T.; Kim, C.-J.; Yi, K. Model Predictive Control of a Semi-Active Suspension with a Shift Delay Compensation Using Preview Road Information. Control Eng. Pract. 2023, 137, 105584. [Google Scholar] [CrossRef]
  23. Wu, K.; Ren, C.; Nan, Y.; Li, L.; Yuan, S.; Shao, S.; Sun, Z. Experimental Research on Vehicle Active Suspension Based on Time-Delay Control. Int. J. Control 2023, 96, 1–17. [Google Scholar] [CrossRef]
  24. Li, G.; Huang, Q.; Hu, G.; Ding, R.; Zhu, W.; Zeng, L. Semi-Active Fuzzy Cooperative Control of Vehicle Suspension with a Magnetorheological Damper. J. Intell. Mater. Syst. Struct. 2023, 1045389X231157353. [Google Scholar] [CrossRef]
  25. Wang, D. Adaptive Control for the Nonlinear Suspension Systems with Stochastic Disturbances and Unknown Time Delay. Syst. Sci. Control Eng. 2022, 10, 208–217. [Google Scholar] [CrossRef]
  26. Zhang, Z.; Dong, J. A New Optimization Control Policy for Fuzzy Vehicle Suspension Systems Under Membership Functions Online Learning. IEEE Trans. Syst. Man Cybern. Syst. 2023, 53, 3255–3266. [Google Scholar] [CrossRef]
  27. Xie, Z.; You, W.; Wong, P.K.; Li, W.; Ma, X.; Zhao, J. Robust Fuzzy Fault Tolerant Control for Nonlinear Active Suspension Systems via Adaptive Hybrid Triggered Scheme. Int. J. Adapt. Control Signal Process. 2023, 37, 1608–1627. [Google Scholar] [CrossRef]
  28. Sakthivel, R.; Shobana, N.; Priyanka, S.; Kwon, O.M. State Observer-Based Predictive Proportional-Integral Tracking Control for Fuzzy Input Time-Delay Systems. Int. J. Robust Nonlinear Control 2023, 33, 6052–6069. [Google Scholar] [CrossRef]
  29. Gu, B.; Cong, J.; Zhao, J.; Chen, H.; Fatemi Golshan, M. A Novel Robust Finite Time Control Approach for a Nonlinear Disturbed Quarter-Vehicle Suspension System with Time Delay Actuation. Automatika 2022, 63, 627–639. [Google Scholar] [CrossRef]
  30. Ma, X.; Wong, P.K.; Li, W.; Zhao, J.; Ghadikolaei, M.A.; Xie, Z. Multi-Objective H-2/H-8 Control of Uncertain Active Suspension Systems with Interval Time-Varying Delay. Proc. Inst. Mech. Eng. Part I J. Syst. Control Eng. 2023, 237, 335–347. [Google Scholar] [CrossRef]
  31. Lee, Y.J.; Pae, D.S.; Choi, H.D.; Lim, M.T. Sampled-Data L-2 - L-8 Filter-Based Fuzzy Control for Active Suspensions. IEEE Access 2023, 11, 21068–21080. [Google Scholar] [CrossRef]
  32. Ma, G.; Wang, Z.; Yuan, Z.; Wang, X.; Yuan, B.; Tao, D. A Comprehensive Survey of Data Augmentation in Visual Reinforcement Learning. arXiv 2022. [Google Scholar] [CrossRef]
  33. Gao, Z.; Yan, X.; Gao, F.; He, L. Driver-like Decision-Making Method for Vehicle Longitudinal Autonomous Driving Based on Deep Reinforcement Learning. Proc. Inst. Mech. Eng. Part D J. Automob. Eng. 2022, 236, 3060–3070. [Google Scholar] [CrossRef]
  34. Fares, A.; Bani Younes, A. Online Reinforcement Learning-Based Control of an Active Suspension System Using the Actor Critic Approach. Appl. Sci. 2020, 10, 8060. [Google Scholar] [CrossRef]
  35. Liu, M.; Li, Y.; Rong, X.; Zhang, S.; Yin, Y. Semi-Active Suspension Control Based on Deep Reinforcement Learning. IEEE Access 2020, 8, 9978–9986. [Google Scholar] [CrossRef]
  36. Pang, H.; Luo, J.; Wang, M.; Wang, L. A Stability Guaranteed Nonfragile Fault-Tolerant Control Approach for Markov-Type Vehicle Active Suspension System Subject to Faults and Disturbances. J. Vib. Control 2023, 10775463231160807. [Google Scholar] [CrossRef]
  37. Kozek, M.; Smoter, A.; Lalik, K. Neural-Assisted Synthesis of a Linear Quadratic Controller for Applications in Active Suspension Systems of Wheeled Vehicles. Energies 2023, 16, 1677. [Google Scholar] [CrossRef]
  38. Li, Y.; Wang, T.; Liu, W.; Tong, S. Neural Network Adaptive Output-Feedback Optimal Control for Active Suspension Systems. IEEE Trans. Syst. Man Cybern Syst. 2022, 52, 4021–4032. [Google Scholar] [CrossRef]
  39. Lin, Y.-C.; Nguyen, H.L.T.; Yang, J.-F.; Chiou, H.-J. A Reinforcement Learning Backstepping-Based Control Design for a Full Vehicle Active Macpherson Suspension System. IET Control Theory Appl. 2022, 16, 1417–1430. [Google Scholar] [CrossRef]
  40. Yong, H.; Seo, J.; Kim, J.; Kim, M.; Choi, J. Suspension Control Strategies Using Switched Soft Actor-Critic Models for Real Roads. IEEE Trans. Ind. Electron. 2023, 70, 824–832. [Google Scholar] [CrossRef]
  41. Lee, D.; Jin, S.; Lee, C. Deep Reinforcement Learning of Semi-Active Suspension Controller for Vehicle Ride Comfort. IEEE Trans. Veh. Technol. 2023, 72, 327–339. [Google Scholar] [CrossRef]
  42. Du, Y.; Chen, J.; Zhao, C.; Liao, F.; Zhu, M. A Hierarchical Framework for Improving Ride Comfort of Autonomous Vehicles via Deep Reinforcement Learning with External Knowledge. Comput.-Aided Civ. Infrastruct. Eng. 2022, 38, 1059–1078. [Google Scholar] [CrossRef]
  43. Han, S.-Y.; Liang, T. Reinforcement-Learning-Based Vibration Control for a Vehicle Semi-Active Suspension System via the PPO Approach. Appl. Sci. 2022, 12, 3078. [Google Scholar] [CrossRef]
  44. Dridi, I.; Hamza, A.; Ben Yahia, N. A New Approach to Controlling an Active Suspension System Based on Reinforcement Learning. Adv. Mech. Eng. 2023, 15, 16878132231180480. [Google Scholar] [CrossRef]
  45. Kwok, N.M.; Ha, Q.P.; Nguyen, T.H.; Li, J.; Samali, B. A Novel Hysteretic Model for Magnetorheological Fluid Dampers and Parameter Identification Using Particle Swarm Optimization. Sens. Actuators A Phys. 2006, 132, 441–451. [Google Scholar] [CrossRef]
  46. Krauze, P.; Kasprzyk, J. Driving Safety Improved with Control of Magnetorheological Dampers in Vehicle Suspension. Appl. Sci. 2020, 10, 8892. [Google Scholar] [CrossRef]
  47. Savaresi, S.M.; Spelta, C. Mixed Sky-Hook and ADD: Approaching the Filtering Limits of a Semi-Active Suspension. J. Dyn. Syst. Meas. Control 2006, 129, 382–392. [Google Scholar] [CrossRef]
  48. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous Control with Deep Reinforcement Learning. arXiv 2019. [Google Scholar] [CrossRef]
  49. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-Level Control through Deep Reinforcement Learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef]
  50. Van Hasselt, H.; Guez, A.; Silver, D. Deep Reinforcement Learning with Double Q-Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. [Google Scholar] [CrossRef]
  51. Fujimoto, S.; Hoof, H.; Meger, D. Addressing Function Approximation Error in Actor-Critic Methods. In Proceedings of the Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 3 July 2018; pp. 1587–1596. [Google Scholar] [CrossRef]
  52. Theunissen, J.; Sorniotti, A.; Gruber, P.; Fallah, S.; Ricco, M.; Kvasnica, M.; Dhaens, M. Regionless Explicit Model Predictive Control of Active Suspension Systems with Preview. IEEE Trans. Ind. Electron. 2020, 67, 4877–4888. [Google Scholar] [CrossRef]
  53. Liang, G.; Zhao, T.; Wei, Y. DDPG Based Self-Learning Active and Model-Constrained Semi-Active Suspension Control. In Proceedings of the 2021 5th CAA International Conference on Vehicular Control and Intelligence (CVCI), Tianjin, China, 29–31 October 2021; pp. 1–6. [Google Scholar] [CrossRef]
  54. Baek, S.; Baek, J.; Choi, J.; Han, S. A Reinforcement Learning-Based Adaptive Time-Delay Control and Its Application to Robot Manipulators. In Proceedings of the 2022 American Control Conference (ACC), Atlanta, GA, USA, 8–10 June 2022; pp. 2722–2729. [Google Scholar] [CrossRef]
  55. Li, S.; Ding, L.; Gao, H.; Liu, Y.-J.; Li, N.; Deng, Z. Reinforcement Learning Neural Network-Based Adaptive Control for State and Input Time-Delayed Wheeled Mobile Robots. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 4171–4182. [Google Scholar] [CrossRef]
  56. Zhu, W.; Garg, T.; Raza, S.; Lalar, S.; Barak, D.D.; Rahmani, A.W. Application Research of Time Delay System Control in Mobile Sensor Networks Based on Deep Reinforcement Learning. Wirel. Commun. Mob. Comput. 2022, 2022, 7844719. [Google Scholar] [CrossRef]
  57. Chen, B.; Xu, M.; Li, L.; Zhao, D. Delay-Aware Model-Based Reinforcement Learning for Continuous Control. Neurocomputing 2021, 450, 119–128. [Google Scholar] [CrossRef]
  58. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal Policy Optimization Algorithms. arXiv 2017. [Google Scholar] [CrossRef]
Figure 1. Dynamics model considering delay time.
Figure 1. Dynamics model considering delay time.
Sensors 23 07827 g001
Figure 2. Road time domain unevenness curve.
Figure 2. Road time domain unevenness curve.
Sensors 23 07827 g002
Figure 3. RL basic framework.
Figure 3. RL basic framework.
Sensors 23 07827 g003
Figure 4. TD3 algorithm improvement structure.
Figure 4. TD3 algorithm improvement structure.
Sensors 23 07827 g004
Figure 5. Critic and actor network. (a) Network architecture created for the critic, and (b) network architecture created for the actor. It should be noted that the last fully connected layer in the critic network directly outputs the result without the need for activation.
Figure 5. Critic and actor network. (a) Network architecture created for the critic, and (b) network architecture created for the actor. It should be noted that the last fully connected layer in the critic network directly outputs the result without the need for activation.
Sensors 23 07827 g005
Figure 6. Control performance with deterministic time delay. (ac) Vehicle body acceleration curves under 10 ms, 20 ms, and 30 ms delay, respectively. (d) Frequency response of x ¨ b / w .
Figure 6. Control performance with deterministic time delay. (ac) Vehicle body acceleration curves under 10 ms, 20 ms, and 30 ms delay, respectively. (d) Frequency response of x ¨ b / w .
Sensors 23 07827 g006
Figure 7. Control performance with semi-regular time delay. (a) Vehicle body acceleration curve under semi-regular time delay. (b) Frequency response of x ¨ b / w .
Figure 7. Control performance with semi-regular time delay. (a) Vehicle body acceleration curve under semi-regular time delay. (b) Frequency response of x ¨ b / w .
Sensors 23 07827 g007
Figure 8. Control performance with uncertain time delay. (a) Vehicle body acceleration curve under uncertain time delay. (b) Frequency response of x ¨ b / w .
Figure 8. Control performance with uncertain time delay. (a) Vehicle body acceleration curve under uncertain time delay. (b) Frequency response of x ¨ b / w .
Sensors 23 07827 g008
Table 1. Environment model parameters.
Table 1. Environment model parameters.
ParametersValueParametersValue
m b 400 m u 40
k b 20,000 k u 200,000
v 20 τ 0
S q 256 c u 0
c b ( active ) 0 c b ( passive ) 1500
Table 2. Agent hyperparameter.
Table 2. Agent hyperparameter.
Hyperparameter
ItemValue
CriticLearning rate1 × 10−3
Gradient threshold1
L2 Regularization factor1 × 10−4
ActorLearning rate1 × 10−2
Gradient threshold1
AgentSample time0.01
Target smoothing factor1 × 10−3
Experience buffer length1 × 106
Discount factor0.99
Mini-batch size128
Soft update factor1 × 10−3
Delayed update frequency2
Noise clipping0.5
Noise variance0.6
The decay rate of noise variance1 × 10−5
Training processMax episodes2000
Max steps1000
Table 3. RMS values of body acceleration in deterministic delayed conditions.
Table 3. RMS values of body acceleration in deterministic delayed conditions.
ControllerPassive10 ms20 ms30 ms
ProposedDDPGProposedDDPGProposedDDPG
x ¨ b ( m / s 2 ) 1.87781.0595
(+43.58%)
0.7772
(+58.61%)
1.0347
(+44.9%)
1.3901
(+25.97%)
1.2716
(+32.28%)
1.5439
(+17.78%)
Table 4. RMS values of body acceleration at different speeds.
Table 4. RMS values of body acceleration at different speeds.
Speed (m/s)PassiveProposedOptimization
101.2650.885330.02%
151.54610.988536.06%
201.78031.111637.56%
251.98371.204939.26%
302.16451.279440.89%
352.32751.35341.87%
402.4761.412242.96%
452.61231.470543.71%
502.73831.517544.58%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Wang, C.; Zhao, S.; Guo, K. Research on Deep Reinforcement Learning Control Algorithm for Active Suspension Considering Uncertain Time Delay. Sensors 2023, 23, 7827. https://doi.org/10.3390/s23187827

AMA Style

Wang Y, Wang C, Zhao S, Guo K. Research on Deep Reinforcement Learning Control Algorithm for Active Suspension Considering Uncertain Time Delay. Sensors. 2023; 23(18):7827. https://doi.org/10.3390/s23187827

Chicago/Turabian Style

Wang, Yang, Cheng Wang, Shijie Zhao, and Konghui Guo. 2023. "Research on Deep Reinforcement Learning Control Algorithm for Active Suspension Considering Uncertain Time Delay" Sensors 23, no. 18: 7827. https://doi.org/10.3390/s23187827

APA Style

Wang, Y., Wang, C., Zhao, S., & Guo, K. (2023). Research on Deep Reinforcement Learning Control Algorithm for Active Suspension Considering Uncertain Time Delay. Sensors, 23(18), 7827. https://doi.org/10.3390/s23187827

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop