Optimal Tuning of PID Parameters Using Iterative

4 downloads 0 Views 545KB Size Report
The importance of PID and challenge in PID auto-tuning at- tracted numerous ..... Table 3 Control performances of G5 −G8 using IMC, PPT and ILT meth- ods.
SICE Journal of Control, Measurement, and System Integration, Vol. 1, No. 2, pp. 143–154, March 2008

Optimal Tuning of PID Parameters Using Iterative Learning Approach Jian-Xin XU ∗ , Deqing HUANG ∗ , and Srinivas PINDI ∗∗ Abstract : Proportional-integral-derivative (PID) controller is the most predominant industrial controller that constitutes more than 90% feedback loops. Time domain performance of PID, including peak overshoot, settling time and rise time, is directly relevant to PID parameters. In this work we propose an iterative learning tuning method (ILT) – an optimal tuning method for PID parameters by means of iterative learning. PID parameters are updated whenever the same control task is repeated. The first novel property of the new tuning method is that the time domain performance or requirements can be incorporated directly into the objective function to be minimized. The second novel property is that the optimal tuning does not require as much the plant model knowledge as other PID tuning methods. The new tuning method is essentially applicable to any plants that are stabilizable by PID controllers. The third novel property is that any existing PID auto-tuning methods can be used to provide the initial setting of PID parameters, and the iterative learning process guarantees that a better PID controller can be achieved. The fourth novel property is that the iterative learning of PID parameters can be applied straightforward to discrete-time or sampled-data systems, in contrast to existing PID auto-tuning methods which are dedicated to continuous-time plants. In this paper, we further exploit efficient searching methods for the optimal tuning of PID parameters. Through theoretical analysis, comprehensive investigations on benchmarking examples, and real-time experiments on the level control of a coupled-tank system, the effectiveness of the proposed method is validated. Key Words : PID tuning, iterative learning, objective functions, searching methods, iterative learning tuning (ILT).

1. Introduction Among all the known controllers, the proportional-integralderivative (PID) controllers are always the first choice for industrial control processes owing to the simple structure, robust performance, and balanced control functionality under a wide range of operating conditions. However, the exact workings and mathematics behind PID methods vary with different industrial users. Tuning PID parameters (gains) remains a challenging issue and directly determines the effectiveness of PID controllers [1]–[3]. To address the PID design issue, much effort has been invested in developing systematic auto-tuning methods. These methods can be divided into three categories, where the classification is based on the availability of a plant model and model type, (i) non-parametric model methods; (ii) parametric model methods; (iii) model free methods. The non-parametric model methods use partial modelling information, usually including the steady state model and critical frequency points. These methods are more suitable for closedloop tuning and applied without the need for extensive priori plant information [1]. Relay feedback tuning method [4]–[7] is a representative method of the first category. The parametric model methods require a linear model of the plant – either transfer function matrix or state space model. To obtain such a model, standard off-line or on-line identification methods are often employed to acquire the model data. Thus parametric model methods are more suitable for off-line PID ∗ ∗∗

Department of Electrical and Computer Engineering, National University of Singapore, Singapore 117576. Department of Electrical Engineering, Indian Institute of Technology Guwahati, Guwahati-781039, India. E-mail: [email protected] (Received November 27, 2007)

tuning [8]. When the plant model is known with a parametric structure, optimal design methods can be applied [9],[10]. As for PID parameter tuning, it can be formulated as the minimization of an objective function with possible design specifications such as the nominal performance, minimum input energy, robust stability, operational constraints, etc. In model free methods, no model or any particular points of the plant are identified. Three typical tuning methods are unfalsified control [11], iterative feedback tuning [12] and extreme seeking [13]. In [11], input-output data is used to determine whether a set of PID parameters meets performance specifications and these PID parameters are updated by an adaptive law based on whether or not the controller falsifies a given criterion. In [12], the PID controller is updated through minimizing an objective function that evaluate the closed-loop performance and estimating the system gradient. In [13], adaptive updating is conducted to tune PID parameters such that the output of the cost function reaches a local minimum or local maximum. In practice, when the plant model is partially unknown, it would be difficult to compute PID parameters even if the relationship between transient specifications and PID parameters can be derived. In many of existing PID tuning methods, whether model free or model based, test signals will have to be injected into the plant in order to find certain relevant information for setting controller parameters. This testing process may however be unacceptable in many real-time control tasks. On the other hand, many control tasks are carried out repeatedly, such as in batch processors. The first objective of this work is to explore the possibility of fully utilizing the task repetitiveness property, consequently provide a learning approach to improve PID controllers through iteratively tuning parameters when the transient behavior is of the main concern.

c 2007 SICE JCMSI 0002/08/0102–0143 

144

SICE JCMSI, Vol. 1, No. 2, March 2008

In most learning methods including neural learning and iterative learning [14], the process Jacobian or gradient plays the key role by providing the greatest descending direction for the learning mechanism to update inputs. The convergence property of these learning methods is solely dependent on the availability of the current information on the gradient. The gradient between the transient control specifications and PID parameters, however, may not be available if the plant model is unknown or partially unknown. Further, the gradient is a function of PID parameters, thus the magnitude and even the sign may vary. The most difficult scenario is when we do not know the sign changes a priori. In such circumstances, traditional learning methods cannot achieve learning convergence. The second objective of this work is to extend the iterative learning approach to deal with the unknown gradient problem for PID parameter tuning. It should be noted that in many industrial control problems such as in process industry, the plant is stable in a wide operation range under closed-loop PID, and the major concern for a PID tuning is the transient behaviors either in the time domain, such as peak overshoot, rise time, settling time, or in the frequency domain such as bandwidth, damping ratio and undamped natural frequency. From the control engineering point of view, it is one of the most challenges to directly address the transient performance, in comparison with the stability issues, by means of tuning control parameters. Even for a lower order linear time invariant plant under PID, the transient performance indices such as overshoot could be highly nonlinear in PID parameters and an analytical inverse mapping from overshoot to PID parameters may not exist. In other words, from the control specification on overshoot we are unable to decide the PID parameters analytically. The third objective of this work is to link these transient specifications with PID parameters and give a systematic tuning method. Another issue is concerned with the redundancy in PID parameter tuning when only one or two transient specifications are required, for instance when only overshoot is specified, or only the integrated absolute error is to be minimized. In order to fully utilize the extra degrees of freedom of the controller, the most common approach is to introduce an objective function and optimize the PID parameters accordingly. This traditional approach is however not directly applicable because of the unknown plant model, and in particular the unknown varying gradient. A solution to this problem is still iterative learning. An objective function, which is accessible, is chosen as the first step for PID parameter optimization. Since the goal is to minimize the objective function, the control inputs will be updated along the greatest descending direction, namely the gradient, of the objective function. In other words, the PID parameters are chosen to directly reduce the objective function, and the objective function is treated as the plant output and used to update the PID parameters. When the gradient is varying and unknown, extra learning trials will be conducted to search the best descending direction. The paper is organized as follows. Section 2 gives the formulation of PID auto-tuning problem. Section 3 introduces the iterative learning tuning approach. Section 4 shows the comparative studies on benchmark examples. Furthermore, Section 5 addresses the real-time implementation on a laboratory pilot plant. Section 6 concludes the work.

2.

Formulation of PID Auto-Tuning Problem

2.1 PID Auto-Tuning In this work, we consider a fundamental PID controller in continuous or discrete-time C(s) = k p + ki

1 + kd s s

C(z) = k p + ki

T sz z−1 + kd z−1 T sz

where k p is the proportional gain, ki the integral gain, kd is the derivative gain, s is the Laplace operator, z is the Z operator, T s is the sampling period. Denote k = [k p , ki , kd ]T . The aim of PID tuning is to find appropriate values for PID parameters such that the closed-loop response can be significantly improved when comparing with the open-loop response. Since the PID control performance is determined by PID parameters k, by choosing a set of performance indices x, for instance overshoot and settling time, there exists a unique relationship or mapping f between x and k x = f(k). The PID auto-tuning problem can be mathematically formulated as to look for a set of k such that x meet the control requirements specified by xd . If the inverse mapping is available, we have k = f −1 (xd ). The mapping f, however, is a vector valued function of x, k and the plant, and is in general a highly nonlinear mapping. Thus its inverse mapping in general is either not analytic solvable or not uniquely existing. Above all, the most difficult problem in PID auto-tuning is the lack of plant model, hence the mapping f is unknown or partially unknown. The importance of PID and challenge in PID auto-tuning attracted numerous researchers and come up with various autotuning or detuning methods, each has unique advantages and limitations. In this work we propose a new auto-tuning method using iterative learning and optimization, which complements existing PID tuning methods. 2.2 Performance Requirements and Objective Functions An objective function, or cost function, quantifies the effectiveness of a given controller in terms of the closed-loop response, either in time domain or frequency domain. A widely used objective function in PID auto-tuning is the integrated square error (ISE) function  T 1 J(k) = e2 (t, k)dt, (1) T − t0 t0 where the error e(t, k) = r(t) − y(t, k) is the difference between the reference, r(t), and the output signal of the closed-loop system, y(t). T and t0 , with 0 ≤ t0 < T < ∞, are two design parameters. In several auto-tuning methods such as IFT and ES, t0 is set approximately at the time when the step response of the closed-loop system reaches the first peak. Hence the cost function effectively places zero weighting on the initial transient portion of the response and the controller is tuned to minimize the error beyond the peak time. Similar objective functions, such as integrated absolute error, integrated time weighted absolute error, integrated time weighted square error,  T  T  T 1 1 1 |e|dt, t|e|dt, te2 dt, T − t0 t0 T − t0 t0 T − t0 t0

SICE JCMSI, Vol. 1, No. 2, March 2008

Fig. 1 The nonlinear mapping between the peak overshoot 100M p and PD gains (k p , kd ) in continuous-time.

145

Fig. 2 The nonlinear mapping between the settling time t s and PD gains (k p , kd ) in continuous-time.

have also been widely used in the process of PID parameter tuning. In many control applications, however, the transient performance, such as overshoot M p , settling time t s , rise time tr , could be of the main concern. An objective function that can capture the transient response directly, is highly desirable. For this purpose, we propose a quadratic function J = (xd − x)T Q(xd − x) + kT Rk,

(2)

where x = [100M p , t s , tr ]T , xd is the desired x, Q and R are two constant non-negative weighting matrices. The selection of Q and R matrices will yield effects on the optimization results, that is, produce different closed-loop responses. A larger Q highlights more on the transient performance, whereas a larger R gives more control penalty. 2.3 A Second Order Example Consider a second order plant under a unity feedback with a PD controller. The PD controller and plant are respectively C(s) = k p + kd s,

k , G(s) = 2 s + as + b

(3)

where k p , kd , k, a, b are all non-negative constants. The closedloop system is stable if a + d > 0 and b + p > 0, where p = kk p and d = kkd . The nonlinear mapping f between the closed-loop transient response x = [M p , t s ] and the PD parameters k = [k p , kd ] is derived in Appendix, and shown in Figs. 1 and 2, where a = 0.1, b = 0, and k = 1. It can be seen that the mapping f is nonlinear or even discontinuous. The nonlinear mapping f for discrete-time control system can also be derived but omitted here due to the complexity. Discretizing the plant in (3) with a sampling time T s = 0.1 s, the nonlinear mapping between M p and (k p , kd ) is shown in Fig. 3. It can be seen that there exist local minima in the surface, and the gradient may vary and take either positive or negative values. On the other hand, it can also been seen from those figures that the transient responses may vary drastically while the control parameters only vary slightly. This indicates the impor-

Fig. 3 The nonlinear mapping between the peak overshoot 100M p and PD gains (k p , kd ) in discrete-time. An upper bound of 100 is applied to crop the vertical values.

tance for PID parameter auto-tuning and the necessity for finding an effective tuning method.

3. Iterative Learning Approach Iterative learning is adopted to provide a solution to the PID auto-tuning. The iterative learning offers a desirable feature that it can guarantee the learning convergence even if the plant model is partially unknown. 3.1 Principal Idea of Iterative Learning The concept of iterative learning was first introduced in control to deal with a repeated control task without requiring the perfect knowledge such as the plant model or parameters [15]. It learns to generate a control action directly instead of doing a model identification. The iterative learning mechanism updates the present control action using information obtained from previous control actions and previous error signals. Let us first give the basic formulation of iterative learning in PID auto-tuning. From preceding discussions, the PID autotuning problem can be described by the mapping x = f(k)

146

SICE JCMSI, Vol. 1, No. 2, March 2008

where k ∈ Ωk ⊂ Rn and x ∈ Ω x ⊂ Rm , where n and m are integer numbers. The learning objective is to find a suitable set k such that the transient response x can reach a given region around the control specifications xd . The principal idea of iterative learning is to construct a contractive mapping A xd − xi+1 = A(xd − xi ),

min J(x, k)

k∈Ωk

where the norm of A is strictly less than 1, and the subscript i indicates that the quantity is in the ith iteration or learning trial. To achieve this contractive mapping, a simple iterative learning law is ki+1 = ki + Γi (xd − xi )

(4)

where Γi ∈ Rn×m is a learning gain matrix. It can be seen that the learning law (4) generates a set of updated parameters from the previously tuned parameters, ki , and previous performance deviations xd −xi . The schematic of the iterative learning mechanism for PID auto-tuning is shown in Fig. 4. When n = m, define the process gradient ∂f(k) , ∂k we can derive the condition for the contractive mapping A F(k) =

xd − xi+1 = xd − xi − (xi+1 − xi ) ∂f(k∗i ) = xd − xi − (ki+1 − ki ) ∂k ∗ = [I − F(ki )Γi ](xd − xi )

where x = f(k). The optimal parameters kopt can be obtained by differentiating J and computing dJ ∂J dx ∂J ∂J ∂J = + = F+ = 0. dk ∂x dk ∂k ∂x ∂k

∈ [min{ki , ki+1 }, max{ki , ki+1 }] ⊂ Ωk . Therefore we where have a contractive mapping A xd − xi+1 = A(xd − xi ) as far as the magnitude (6)

When n > m, there exists an infinite number of solutions because of redundancy in control parameters. With the extra

(7)

In most cases kopt can only be found numerically due to the highly nonlinear relationship between k and quantities J, x. A major limitation of optimization methods is the demand for the complete plant model knowledge or the mapping f. On the contrary, iterative learning only requires the bounding knowledge of the process gradient. The principal idea of iterative learning can be extended to solving the optimization problem under the assumption that all gradient components with respect to k have known limiting bounds. Since the learning objective now is to directly reduce the value of J, the objective function J can be regarded as the process output to be minimized. The new iterative learning tuning law is ki+1 = ki − γi J(ki ),

(5)

k∗i

|A| = |I − F(k∗i )Γi | ≤ ρ < 1.

degrees of freedom in PID, optimality can be exploited, for instance the shortest settling time, minimum peak overshoot, minimum values of control parameters, etc. A suitable objective function to be minimized could be a non-negative function J(x, k) ≥ 0, where J is accessible, such as the quadratic one (2). The minimization is in fact a searching task

(8)

where γi = [γ1,i , · · · , γn,i ]T and J(ki ) denotes J(ki , f(ki )). To show the contractive mapping, note that J(ki+1 ) = J(ki ) + [J(ki+1 ) − J(ki )]   dJ(k∗i ) T (ki+1 − ki ) = J(ki ) + dk where k∗i ∈ Ωk is in a region specified by ki and ki+1 . By substitution of the ILT law (8), we have ⎤ ⎡   dJ(k∗i ) T ⎥⎥⎥ ⎢⎢⎢ γi ⎦⎥ J(ki ). (9) J(ki+1 ) = ⎣⎢1 − dk The convergence property is determined by the learning gains γi and the gradient dJ/dk. 3.2 Learning Gain Design Based on Gradient Information To guarantee the contractive mapping (9), the magnitude relationship must satisfy |1 − DTi γi | ≤ ρ < 1, where   dJ(k∗i ) Di = dk

Fig. 4 The schematic block diagram of the iterative learning mechanism and PID control loop. The parameter correction is generated by the performance deviations xd − xi multiplied by a learning gain Γi . The operator z−1 denotes one iteration delay. The new PID parameters ki+1 consists of the previous ki and the correction term, analogous to a discrete-time integrator. The iterative learning tuning mechanism is shown by the block enclosed by the dashed line. r is the desired output and the block M is a feature extraction mechanism that records the required transient quantities such as overshoot from the output response yi+1 .

is the gradient of the objective function. The selection of learning gain γi is highly related to the prior knowledge on the gradient Di . Consider three scenarios and assume Di = (D1,i , D2,i , D3,i ), that is, a PID controller is used. When Di is known a priori, we can choose γ j,i = D−1 j,i /3. Such a selection produces the fastest learning convergence speed, that is, convergence in one iteration because 1−DTi γi = 0. When the bounding knowledge and the sign information of D j,i are available, the learning convergence can also be guaranteed. For instance assume 0 < α j ≤ D j,i ≤ β j < ∞ for

SICE JCMSI, Vol. 1, No. 2, March 2008

k∗i ∈ Ωk , where α j and β j are respectively lower and upper bounds of the gradient components D j,i . In such circumstances, choosing γ j,i = 1/3β j , the upper bound of the magnitude is ρ = 1 − α1 /3β1 − α2 /3β2 − α3 /3β3 < 1. The most difficult scenario is when bounding functions or signs of D j,i are unknown. In order to derive the iterative learning convergence, it can be seen from (9) that we do not need the exact knowledge about the mapping f. It is adequate to know the bounding knowledge (amplitude and sign) of the gradient D(k∗i ) or F(k∗i ). Although an invariant gradient is assumed for most iterative learning problems, the elements in D(k∗i ) may change sign and take either positive or negative values. Without knowing the exact knowledge of the mapping f, we may not be able to predict the varying signs in the gradient. Now, in the iterative learning law (8) the learning gains will have to change signs according to the gradient. Here the question is how to change the signs of the learning gains when we do not know the signs of the gradient components, i.e. how to search the direction of the gradient, which obviously can only be done in a model free manner if f is unknown. A solution to the problem with unknown gradient is to conduct extra learning trials to determine the direction of gradient or the signs of the learning gains directly γi = [±γ1 , · · · , ±γn ]T ,

147

Fig. 5

There are four pairs of signs for the gradient (D1 , D2 ) as indicated by the arrows. Hence there are four possible updating directions, in which one pair gives the fastest descending direction.

Fig. 6

There are three gradient components D1 , D2 and D3 with respect to three control parameters. Consequently there are 8 possible tuning directions and at most 8 learning trials are required to find the correct updating direction.

(10)

where γi are positive constants. From the derivation (5), when learning gains are chosen appropriately, |A| < 1 and the learning error reduces. On the other hand, if learning gains are chosen inappropriately, then |A| > 1 and the error increases after this learning trial. Therefore, several learning trials are adequate for ILT mechanism to determine the correct signs of the learning gains. In general, when there are two gradient components (D1,i , D2,i ), there are 4 sets of signs {1, 1}, {1, −1}, {−1, 1}, and {−1, −1}, corresponding to all possibles signs of the gradient (D1,i , D2,i ). In such circumstances, at most 4 learning trials are sufficient to find the greatest descending among the four control directions, as shown in Fig. 5. Similarly, if there are three free tuning parameters, there will be 8 sets of signs in the gradient (D1,i , D2,i , D3,i ), as shown in Fig. 6. In general, if there are n control tuning parameters, there will be n gradient components D j,i . Since each γ j,i takes either positive or negative sign, there will be 2n combinations and at most 2n learning trials are required. In addition to the estimation of the gradient direction or learning direction, the magnitudes of learning gains γi should also be adjusted to satisfy the learning convergence condition |1 − DTi γi | ≤ ρ < 1. Since the gradient Di is a function of PID parameters ki , the magnitude of Di varies at different iterations. When the magnitudes of the gradient is unknown, extra learning trials will be needed to search for suitable magnitudes of learning gains. In this work, we adopt a self-adaption rule [16] which scales learning gains up and down by a factor ζ = 1.839, that is, each component γ j,i will be adjusted to γ j,i ζ and γ j,i /ζ. Extra learning trials will be performed with the scaled learning gains. It is reported that the choice of such a scaling factor ζ will lead to a linear convergence rate [16],[17]. To facilitate searching and reduce the number of trials, we can also estimate gradient components D j,i numerically using

the values of the objective function and PID parameters obtained from previous two iterations J(ki−1 ) − J(ki−2 ) Dˆ j,i = . k j,i−1 − k j,i−2

(11)

The learning gain can be revised accordingly as γ j,i = λ j Dˆ −1 j,i where λ j is a constant gain in the interval of (0, 1]. This method is in essence the Secant method along the iteration axis, which can effectively expedite the learning speed [18]. Since gradient direction is a critical issue in searching, and the approximation (11) may not always guarantee a correct sign, we can use the estimation result in (11) partially by retaining the magnitude estimation, while still searching the correct control direction k j,i−1 − k j,i−2 γ j,i = ±λ j . (12) J(ki−1 ) − J(ki−2 ) 3.3 Iterative Searching Methods Three iterative searching methods are considered for ILT in this work with verifications and comparisons. They are M0 – an exhaustive searching method in directions and magnitude, M1 – an exhaustive searching method in directions, M2 – a lazy searching method. M0 does exhaustive searching in all 2n directions and exhaustive searching in all 2n magnitudes using self-adaptation with the factor ζ. Then PID parameters will be updated using the set that generates the best closed-loop response or yields the

148

SICE JCMSI, Vol. 1, No. 2, March 2008

biggest drop of J in that trial. With the best tuned PID parameters as the initial setting, the ILT mechanism enters another run of exhaustive searching for the best response and best PID parameters. The searching process repeats until the stopping criterion is met. This is the worst case searching where neither the gradient directions nor the gradient magnitudes are available. For each run of searching the greatest descending direction, 4n trials are performed. For PID tuning where n = 3, 64 trials are needed for one run of searching. Clearly, this searching method is not efficient. In M1 the entire searching and updating process is similar to M0 except that the magnitudes of learning gains are determined using the formula (12). Hence it reduces the total number of trials from 4n in M0 to 2n for each run of searching. For PID tuning where n = 3, only 8 trials are needed. M2 does exhaustive searching in directions in the first run of searching, and the magnitudes of learning gains are determined using the formula (12). Then the greatest descending direction will be used for subsequent runs of searching, with the assumption that the mapping f is in general smooth and drastic variations in gradient rarely occur. The exhaustive searching in directions will be activated again when the stopping criterion is met. The searching process will permanently stop if the stop criterion is still met after exhaustive searching in all directions. The initial magnitudes of learning gains can be set as γ0 [γ1,0 , γ2,0 , γ3,0 ] = [k p,0 , ki,0 , kd,0 ], (13) J0 where γ0 is a positive constant, and chosen to be 0.1 in this work. k p,0 , ki,0 , kd,0 are initial values of PID parameters determined using any existing PID auto-tuning methods. J0 is calculated with k p,0 , ki,0 , kd,0 and the corresponding closed-loop response.

4.

Comparative Studies on Benchmark Examples

In this section we conduct comprehensive tests on 8 benchmark plant models. Four plant models G1 to G4 were used in [13] to compare and validate several iterative tuning methods 1 G1 (s) = e−5s , 1 + 20s 1 G2 (s) = e−20s , 1 + 20s 1 , G3 (s) = (1 + 10s)8 1 − 5s G4 (s) = , (1 + 10s)(1 + 20s) where G1 and G2 have relatively large normalized time delay (NTD), G3 has high-order repeated poles, and G4 is nonminimum phase. The other four plant models G5 to G8 were used in [19],[20] to validate their auto-tuning methods 1 e−0.5s , (s + 1)(s + 5)2 1 e−s , G6 (s) = (25s2 + 5s + 1)(5s + 1) 1 e−0.3s , G7 (s) = 2 (s + 2s + 3)(s + 3) 1 G8 (s) = 2 e−0.1s , (s + s + 1)(s + 2)2 G5 (s) =

where G5 is a high-order plant with medium NTD, G6 is a highorder and moderately oscillatory plant with short NTD, G7 is a high-order and heavily oscillatory plant with short NTD, and G8 has both oscillatory and repeated poles. To make fair comparisons, we choose initial learning gains with form (13) for PID parameters in all case studies. In all searching results, we use N to denote the total number of trials. 4.1 Comparisons between Objective Functions Objective function ISE (1) has been widely investigated and adopted in PID tuning. The quadratic objective function (2), on the other hand, has more weights to prioritize transient performance requirements. When control requirements are directly concerned with transient response such as M p or t s , we can only use the quadratic objective function (2). To show the effects of different objective functions and weight selections, we use plants G1 – G4 . To make fair comparisons, the learning process starts from the same initial setting for PID gains generated by Ziegler Nicholes (ZN) tuning method [8]. Exhaustive searching method M0 is employed. The tuning results through iterative learning are summarized in Table 1. For simplicity only M p and t s are taken into consideration in (2). The learning process stops when the drop of an objective function between two consecutive iterations is lower than ε = 10−6 for (1) and ε = 0.01 for (2). In ISE, the parameters are T = 100, 300, 500, 200 s and t0 = 10, 50, 140, 30 s respectively [13]. Usually settling time is much greater than overshoot, thus 100M p is used. When plants have much bigger settling time, we can choose t s instead of (t s )2 in the objective function, as shown in cases of G2 and G3 . By scaling down t s in objective functions, overshoot decreases as it is weighted more. Meanwhile t s increases as it is weighted less. Comparing ISE and quadratic objective function, we can see that latter offers more choices. 4.2 Comparisons between ILT and Existing Iterative Tuning Methods Now we compare iterative learning based tuning with other iterative tuning methods such as extremum seeking (ES) tuning and iterative feedback tuning (IFT). ZN tuning method is also included for comparison. G1 – G4 and the same ISE [13] for each model are used. The PID controller parameters given by ZN are used as a starting point for ILT tuning. Exhaustive searching method M0 is employed in ILT. The results verify that ILT, ES and IFT give very similar responses which are much superior than that of ZN tuning method. Figure 7 shows the ILT results for G1 , (a) shows the decreasing objective function J, (b) shows decreasing performance indices M p and t s , (c) shows the variations of the PID parameters, and (d) compares step responses with four tuning methods. Figure 8 shows the searching results of the gradient directions and the variations of the learning gains through selfadaptation. It can be seen that the gradients undergo changes in signs, hence it is in general a difficult and challenging optimization problem. Table 2 summarizes the comparative results with all four plants G1 – G4 when 4 tuning methods were applied. The iteration numbers when applying ILT for G1 – G4 are 786, 512, 1024 and 1216 respectively. Although ILT shares similar performance as ES and IFT, it

149

SICE JCMSI, Vol. 1, No. 2, March 2008

Table 1 Plant G1

G2

G3

G4

Control performances of G1 − G4 using the proposed ILT method.

J ISE (100M p )2 + 0.5t2s (100M p )2 + 0.1t2s (100M p )2 + 0.01t2s ISE (100M p )2 + t s (100M p )2 + 0.2t s (100M p )2 + 0.1t s ISE (100M p )2 + t s (100M p )2 + 0.2t s (100M p )2 + 0.1t s ISE (100M p )2 + 0.25t2s (100M p )2 + 0.1t2s (100M p )2 + 0.01t2s

PID Controller 3.59 + 0.13s−1 + 7.54s 3.38 + 0.13s−1 + 7.05s 3.25 + 0.12s−1 + 6.30s 3.71 + 0.11s−1 + 9.11s 0.93 + 0.031s−1 + 5.67s 0.99 + 0.032s−1 + 6.87s 1.05 + 0.028s−1 + 9.79s 1.03 + 0.029s−1 + 9.18s 0.64 + 0.012s−1 + 11.3s 0.76 + 0.013s−1 + 16.65s 0.85 + 0.014s−1 + 25.77s 0.83 + 0.014s−1 + 24.91s 5.01 + 0.092s−1 + 25.59s 4.31 + 0.075s−1 + 22.19s 3.89 + 0.071s−1 + 22.28s 4.51 + 0.075s−1 + 23.96s

100M p 4.48 1.99 0.63 0.53 0.71 1.06 0.29 0.20 0.49 1.93 0.66 0.62 3.05 1.81 1.70 0.06

ts 21.35 12.17 12.83 22.24 50.67 47.99 82.74 83.61 137.13 120.56 212.04 212.76 25.2 18.63 20.56 19.27

N 768 1024 832 512 512 1600 512 640 1024 576 192 192 1216 512 384 1216

Fig. 7 ILT performance for G1 . (a) The evolution of the objective function; (b) The evolution of overshoot and settling time; (c) The evolution of PID parameters; (d) The comparisons of step responses among ZN, IFT, ES and ILT, where IFT, ES and ILT show almost the same responses.

is worth to highlight some important factors in the tuning process. In ES tuning, there are more than 10 design parameters to be set properly. From [13], design parameters take rather specific values. In IFT, the initial values of the PID parameters must be chosen in such a way as to give an initial response that is very slow and with no overshoot. Further, in IFT the transient performance is purposely excluded from the objective function by choosing a sufficiently large t0 . On the contrary, in ILT only initial learning gains need to be preset, and we choose all three learning gains with a uniform value γ0 = 0.1 in (13) for the ease of ILT design. In fact, by choosing initial learning gains with different values, we can achieve much better responses than those in Table 2. Needless to mention that ILT can handle both types of objective functions (1) and (2) with the flexibility to highlight transient behaviors.

4.3 Comparisons between ILT and Existing Auto-Tuning Methods PID auto-tuning methods [8],[19],[20] provided several effective ways to determine PID parameters. In this subsection we compare ILT with the auto-tuning method [8] based on internal model control (IMC), and the auto-tuning method [20] based on pole-placement (PPT) which shows superior performance than [19]. Comparisons are made based on plants G5 – G8 which were used as benchmarks. Using ISE as the objective function and searching method M0 in ILT, the tuning results are summarized in Table 3, where the start points of PID parameters are adopted from the tuning results in [19]. Comparing with the results auto-tuned by the IMC method and PPT method, ILT achieves better performance after learning.

150

SICE JCMSI, Vol. 1, No. 2, March 2008

Fig. 8 ILT searching results for G1 . (a) The evolution of the gradient directions; (b) The evolution of the magnitudes of learning gains with self-adaptation. Table 2 Control performances of G1 −G4 using methods ZN, IFT, ES and ILT. Plant Method PID Controller 100M p ts ZN 4.06 + 0.44s−1 + 9.38s 46.50 47.90 G1 IFT 3.67 + 0.13s−1 + 7.74s 5.38 21.38 ES 3.58 + 0.13s−1 + 7.68s 3.31 21.35 ILT 3.59 + 0.13s−1 + 7.54s 4.48 21.35 ZN 1.33 + 0.043s−1 + 10.30s 21.75 109.4 G2 IFT 0.93 + 0.031s−1 + 5.64s 0.80 50.33 ES 1.01 + 0.032s−1 + 7.23s 1.37 76.61 ILT 0.93 + 0.031s−1 + 5.67s 0.71 50.67 ZN 1.10 + 0.015s−1 + 20.91s 13.95 336.90 G3 IFT 0.66 + 0.012s−1 + 12.08s 0.98 132.05 ES 0.68 + 0.013s−1 + 13.30s 0.96 130.41 ILT 0.64 + 0.012s−1 + 11.30s 0.49 137.13 ZN 3.53 + 0.21s−1 + 14.80s 53.70 86.12 G4 IFT 3.03 + 0.065s−1 + 18.42s 0.55 28.74 ES 3.35 + 0.068s−1 + 21.40s 0.18 29.80 ILT 5.01 + 0.092s−1 + 25.59s 3.05 25.21 Table 3 Control performances of G5 − G8 using IMC, PPT and ILT methods. Plant Method PID Controller 100M p ts IMC 33.46 + 16.67s−1 + 8.54s 2.09 5.71 G5 PPT 27.02 + 21.14s−1 + 6.45s 12.12 5.01 ILT 31.96 + 18.81s−1 + 11.01s 1.31 2.09 IMC 0.88 + 0.066s−1 + 1.20s 3.46 74.61 G6 PPT 0.44 + 0.074s−1 + 2.54s 5.10 49.81 ILT 0.57 + 0.075s−1 + 2.78s 1.17 23.91 IMC 7.06 + 5.29s−1 + 1.64s 5.15 8.37 G7 PPT 3.89 + 5.39s−1 + 2.15s 3.21 5.32 ILT 4.51 + 5.40s−1 + 2.53s 0.44 2.93 IMC 2.79 + 1.33s−1 + 1.19s 3.01 14.74 G8 PPT 1.50 + 1.37s−1 + 1.72s 3.04 9.44 ILT 2.18 + 1.48s−1 + 2.34s 1.18 4.61

4.4 Comparisons between Searching Methods Now we investigate the effects of searching methods M0 , M1 and M2 . Plants G1 – G8 are used. The objective function to be used is J = (100M p )2 +

t2s . 100

(14)

Set λ j = 0.2, j = 1, 2, 3 in (12) for M1 and M2 . The stopping criterion for ILT is ε = 0.01. The control performance after

learning is summarized in Tables 4 and 5. It can be seen that M1 and M2 achieve similar performance, which is slightly inferior than M0 . However, comparing with M0 the learning trial numbers in M1 and M2 have been significantly reduced. 4.5 ILT for Sampled-Data Systems A promising feature of ILT is the applicability to sampleddata or discrete-time systems. To illustrate how ILT works for digital systems, consider plant G4 which can be discretized using sampler and zero order hold

−2.5e−.05T s + 1.5e−.1T s + 1 z G4 (z) =

+ e−.15T s − 2.5e−0.1T s + 1.5e−.05T s z2 − (e−0.1T s + e−0.05T s )z + e−0.15T s

,

where the sampling period T s = 0.1 s. The digital PID controller is used. Choose again (14) as the objective function, and use ZN to generate the initial values for PID parameters. The closed-loop responses using ILT are summarized in Table 6. For comparison all three searching methods M0 , M1 and M2 are used. It can be seen that in all cases the control responses have been improved drastically, especially the reduction in overshoot.

5.

Real-Time Implementation

In order to show the applicability and effectiveness of the proposed ILT, real-time experiment has been carried out on a couple tank. 5.1 Experimental Setup and Plant Modelling The couple tank equipment consists of two small perspex tower-type tanks and interconnected through an hole which yields a hydraulic resistance (Fig. 9). The experimental setup in this work was configured such that the level is measured from the tank-2 while the water is pumped into the tank-1 as the control input. The outlet of tank-2 is used to discharge the water into the reservoir. The measurement data is collected from couple tank using NI data acquisition card USB-6008. The control method is programmed using Labview. A window-type smoothing filter using 100 samples is implemented to mitigate measurement noise. The sampling period is 0.125 second.

151

SICE JCMSI, Vol. 1, No. 2, March 2008

Table 4

Control performance of G1 − G8 using searching methods M0 , M1 , M2 .

Plant G1 G2 G3 G4 G5 G6 G7 G8

100M p 0.53 1.80 1.81 0.06 0.282 0.605 0 0.19 Table 5

Plant G1 G2 G3 G4 G5 G6 G7 G8

M0 ts 22.24 78.64 121.29 19.27 2.16 25.04 6.05 8.25

N 512 256 192 1216 640 320 192 768

100M p 0.24 0.73 1.36 0 0.21 0.03 0 0

M1 ts 22.13 79.93 209.77 15.92 2.44 48.04 6.13 10.53

M2 N 160 72 48 120 48 40 32 32

100M p 0.23 1.47 1.81 0 0 0.03 0 0

ts 22.23 79.91 205.35 22.93 3.34 48.04 6.15 10.74

N 29 43 41 41 70 19 28 18

Final controllers for G1 − G8 by using searching methods M0 , M1 , M2 .

M0 3.59 + 0.13s−1 + 7.54s 0.93 + 0.029s−1 + 9.43s 0.75 + 0.013s−1 + 16.26s 4.51 + 0.075s−1 + 23.96s 31.52 + 18.31s−1 + 10.76s 0.51 + 0.070s−1 + 2.14s 5.34 + 4.86s−1 + 1.28s 2.03 + 1.27s−1 + 2.71s

M1 3.42 + 0.11s−1 + 8.77s 0.94 + 0.030s−1 + 7.28s 0.87 + 0.012s−1 + 24.73s 5.99 + 0.097s−1 + 30.70s 29.63 + 17.65s−1 + 10.19s 0.60 + 0.069s−1 + 2.17s 5.02 + 4.81s−1 + 1.31s 2.13 + 1.21s−1 + 0.94s

M2 3.38 + 0.12s−1 + 7.80s 1.05 + 0.031s−1 + 8.98s 0.72 + 0.012s−1 + 21.42s 6.92 + 0.10s−1 + 39.18s 17.00 + 13.68s−1 + 0.54s 0.60 + 0.069s−1 + 2.17s 4.11 + 4.81s−1 + 1.31s 1.74 + 1.21s−1 + 0.94s

Table 6 Digital Control Results. Initial performance is achieved by ZN tuned PID. Final performance is achieved by ILT. Ts

Initial Performance 100M p ts

0.01

39.16

64.14

0.05

39.52

63.90

0.2

40.92

63.00

0.5

44.26

77.00

2

76.96

88.00

Method M0 M1 M2 M0 M1 M2 M0 M1 M2 M0 M1 M2 M0 M1 M2

Final k p , ki , kd 3.56, 0.10, 21.62 4.04, 0.12, 25.26 3.24, 0.11, 21.12 3.45, 0.10, 20.89 4.04, 0.12, 25.26 3.25, 0.11, 21.12 3.53, 0.11, 21.76 4.08, 0.15, 28.21 2.80, 0.090, 16.41 3.39, 0.11, 21.60 3.48, 0.13, 24.05 2.69, 0.086, 15.61 2.22, 0.082, 17.60 2.19, 0.078, 15.22 2.16, 0.075, 13.94

Final Performance 100M p ts N 0.13 14.77 896 0.43 11.25 80 1.12 17.64 43 0.057 15.45 768 0.55 10.85 80 1.10 17.35 43 0.060 13.60 1024 1.02 11.80 96 0.70 20.20 56 0.00 11.50 832 0.97 15.50 72 0.74 20.00 48 1.08 34.00 576 0.32 34.00 128 0.0020 18.00 81

A step response is conducted to approximate the couple tank dynamics with a first order plus time delay model [8] G(s) =

k e−sL τs + 1

(15)

where k is the plant DC gain, τ is the time constant, L is the transportation lag. As shown in Fig. 10, after conducting a step response the obtained plant is G(s) =

168 −3.6s . e 1 + 65s

(16)

5.2 Application of ILT Method In the experiments, we conducts a series of tests to investigate ILT. The quadratic objective function is J = (100M p )2 + qt2s

Fig. 9

Diagram of couple tank apparatus.

where two values q = 0.1 and q = 0.01 are used. The ISE is  T 1 e2 dt, T − t0 t0

152

SICE JCMSI, Vol. 1, No. 2, March 2008

Fig. 10 Step response based modelling. Table 7 Experimental results. Method M0

M1

M2

Constraint C1 C1 C2 C1 C2 C1 C2 C1 C2 C1 C2 C1 C2

J (100M p )2 + 0.01t2s ISE ISE (100M p )2 + +0.01t2s (100M p )2 + 0.01t2s (100M p )2 + 0.1t2s (100M p )2 + 0.1t2s ISE ISE (100M p )2 + 0.01t2s (100M p )2 + 0.01t2s (100M p )2 + 0.1t2s (100M p )2 + 0.1t2s

where T = 124 s and t0 = 27.4 s. All three search methods M0 , M1 and M2 are applied. It is observed in the experiment that the calculated gradient components (11) or (12) may be singular some times. This is due to the presence of measurement noise and the closeness of the values of J at two adjacent iterations. On the other hand, the learning process may become sluggish when the PID parameters at two adjacent iterations are too close, yielding a very lower learning gains γ j,i . To solve these two problems, a constraint is applied when updating the learning gains c1 |k j,i | ≤ γ j,i ≤ c2 |k j,i |,

(17)

where 0 ≤ c1 < c2 ≤ 1. The upper and lower learning gain bounds are made in proportion to PID parameters. The rationale is clear. When a controller parameter is bigger, we can update it with a larger bound without incurring drastic changes. If setting absolute bounds for the learning gain, an overly small bound would limit the parameter updating speed, and an overly large bound would make the constraint ineffective. In the realtime application we consider 2 sets of boundaries C1 : c1 = 0.02 c2 = 0.2, C2 : c1 = 0.05 c2 = 0.4. The initial PID controller is found using ZN tuning method 0.11 +

0.011 + 0.26s. s

The transient performance with ZN tuned PID is 100M p = 30.02 and t s = 58.73 s. ILT is applied to improve the performance.

PID Controller 0.11 + 0.0029s−1 + 0.26s 0.16 + 0.0045s−1 + 0.32s 0.14 + 0.0039s−1 + 0.20s 0.21 + 0.0036s−1 + 0.43s 0.16 + 0.0041s−1 + 0.34s 0.26 + 0.0037s−1 + 0.45s 0.22 + 0.0036s−1 + 0.44s 0.26 + 0.0040s−1 + 0.62s 0.23 + 0.0042s−1 + 0.56s 0.32 + 0.0034s−1 + 0.76s 0.30 + 0.0034s−1 + 0.72s 0.30 + 0.0038s−1 + 0.72s 0.23 + 0.0037s−1 + 0.56s

100M p 0.089 4.18 4.29 0.11 0.78 1.40 1.44 1.30 2.19 0.27 0.50 1.30 1.71

ts 16.65 34.04 30.22 16.58 10.45 9.35 10.10 10.50 12.98 17.30 16.63 16.33 15.99

N 384 56 32 56 24 48 32 15 13 17 14 14 10

5.3 Experimental Results The results are summarized in Table 7. It can be seen that searching methods M1 and M2 can significantly reduce the number of learning iterations while M p and t s can be maintained at almost the same level. By changing the weight for t s , the final overshoot M p and settling time t s can be adjusted. The constraint C2 can also reduce the trial number, because higher limits in learning gains will expedite the learning progress. In experiments, it is observed that M2 can be further simplified by removing the last run of exhaustive searching, so long as the variation of J reaches the preset threshold ε.

6.

Conclusion

A new PID auto-tuning method is developed and compared with several well established auto-tuning methods including ZN, IFT, ES, PPT, IMC. Iterative learning tuning provides a sustained improvement in closed-loop control performance, and offers extra degrees of freedom in specifying the transient control requirements through a new objective function. The iterative learning tuning method proposed in this work can be further extended to various optimal designs for controllers, owing to its model free nature. For instance, by iterative searching and task repeatability, we can tune the parameters of lead and lag compensators, filters, observers, and use both time domain and frequency performance indices in objective functions. The future issues associated with the iterative learning tuning approach we are looking into include the robustness of ILT when the control process is not completely repeatable, the incorporation of various practical constraints, improvement of searching efficiency for high dimension multi-loop control systems, as well as ILT for closed-loop systems with lower sta-

153

SICE JCMSI, Vol. 1, No. 2, March 2008

bility margins such as the sampled-data system with large time delays. References [1] M. A. Johnson and M. H. Moradi: PID Controller Design, Spring-Verlag, London, 2003. [2] A. O’Dwyer: Handbook of PI and PID Controller Tuning Rules, Imperial College Press, London, 2003. [3] K. J. Astrom, T. Hagglund, C. C. Hang, and W. K. Ho: Automatic tuning and adaptation for PID controllers – A survery, Control Engineering Practice, Vol. 1, No. 4, pp. 699–714, 1993. [4] K. J. Astrom and T. Hagglund: Automatic tuning of simple regulators with specifications on phase and amplitude margins, Automatica, Vol. 20, No. 5, pp. 645–651, 1984. [5] A. A. Voda and I. D. Landau: A method for the auto-calibration of PID controllers, Automatica, Vol. 31, No. 1, pp. 41–53, 1995. [6] A. Leva: PID autotuning method based on relay feedback, IEEE Proceeding D – Control Theory and Applications, Vol. 140, No. 5, pp. 328–338, 1993. [7] G. K. I. Mann, B.-G. Hu, and R. G. Gosine: Time-domain based design and analysis of new PID tuning rules, IEE Proceedings on Control Theory and Applications, Vol. 148, No. 3, pp. 251–261, 2001. [8] K. J. Astrom and T. Hagglund: PID Controllers: Theory, Design and Tuning, 2nd ed., Research Triangle Park, NC: Instrum. Soc. Amer., 1995. [9] M. Ikeda and D. D. Siljak: Optimality and robustness of linearquadratic control for nonlinear systems, Automatica, Vol. 26, No. 3, pp. 499–511, 1990. [10] C. Y. Kuo, C. L. Yang, and C. Margolin: Optimal controller design for nonlinear systems, IEE Proceedings on Control Theory and Applications, Vol. 145, No. 1, pp. 97–105, 1998. [11] M. Saeki: Unfalsified control approach to parameter space design of PID controllers, Proceedings of 42nd IEEE Conference on Decision and Control, pp. 786–791, 2003. [12] O. Lequin, E. Bosmans, and T. Triest: Iterative feedback tuning of PID parameters: Comparison with classical tuning rules, Control Engineering Practice, Vol. 11, No. 9, pp. 1023–1033, 2003. [13] N. J. Killingsworth and M. Krstic: PID tuning using extremum seeking: online, model-free performance optimization, IEEE Control Systems Magazine, Vol. 26, No. 1, pp. 70–79, 2006. [14] Z. Bien and J.-X. Xu: Iterative Learning Control - Analysis, Design, Integration and Applications, Kluwer Academic Press, Boston, USA, 1998. [15] A. Arimoto, S. Kawamura, and F. Miyazaki: Bettering operation of robots by learning, Journal of Robotic Systems, Vol. 1, pp. 123–140, 1984. [16] R. Salomon: Evolutionary methods and Gradient Search: Similarities and Differences, IEEE Transactions On Evolutionary Computation, Vol. 2, No. 2, 1998. [17] R. Salomon and J. L. van Hemmen: Accelerating backpropagation through dynamic self-adaptation, Neural Networks, Vol. 9, No. 4, pp. 589–601, 1996. [18] J.-X. Xu and Y. Tan: Linear and Nonlinear Iterative Learning Control, Springer-Verlag, Berlin, 2003. [19] W. K. Ho, O. P. Gan, E. B. Tay, and E. L. Ang: Performance and gain and phase margins of well-known PID tuning formulas, IEEE Transactions on Control Systems Technology, Vol. 4, No. 3, pp. 473–477, 1996. [20] Q. Wang, T. Lee, H. Fung, Q. Bi, and Y. Zhang: PID tuning for improved performance, IEEE Transactions on Control Systems Technology, Vol. 7, No. 4, pp. 457–465, 1999.

Appendix The nonlinear mapping f will be explored for three cases with underdamped, overdamped and critical damped characteristics. Underdamped Case When the system parameters satisfy (a+kkd )2 −4(b+kk p ) < 0, the closed-loop system shows underdamped response with two complex poles. The system response in the time domain is √ y(t) = A − A 1 + λ2 e−αt cos(βt − tan−1 λ), (A. 1) where p a+d , α= , b+ p 2  1 β = 4(b + p) − (a + d)2 , 2 ap − 2bd − pd λ =  p 4(b + p) − (a + d)2

A =

with p = kk p and d = kkd . It is easy to see that the steady-state output is A and √

M p = − 1 + λ2 e−αt p cos βt p − tan−1 λ , (A. 2)

1 √ t s = ln 50 1 + λ2 , (A. 3) α where the peak time is ⎧ 1 βλ−α ⎪ tan−1 β+αλ ⎪ ⎪ ⎨ β tp = ⎪  ⎪ ⎪ ⎩ 1 π + tan−1 β

 βλ−α

β+αλ

if

βλ−α β+αλ

≥ 0,

if

βλ−α β+αλ

< 0.

Note that the settling time t s , corresponding to 2% of the final steady state value, is calculated from the worst case condition √

1 + λ2 e−αt cos βt − tan−1 λ < 0.02

by ignoring the factor cos βt − tan−1 λ . Overdamped Case When the system parameters satisfy (a + kkd )2 − 4(b + kk p ) > 0, the closed-loop shows overdamped response and the closedloop system has two distinct real poles. The system response in the time domain is   p 1 d − e−(α+β)t y(t) = A − 2 β β(α + β)   1 d p + (A. 4) − e−(α−β)t . 2 β β(α − β) where α=

1 a+d , β= 2 2

 (a + d)2 − 4(b + p).

The relationship α > β always holds since b + p > 0. Therefore the steady state is still A. First derive the peak overshoot. From dy 1 = [(dα + dβ − p)e−(α+β)t + (−dα + dβ + p)e−(α−β)t ], dt 2β and let dy/dt = 0 we can obtain the peak time   dα − dβ − p 1 t p = − ln . 2β dα + dβ − p The output has overshoot if and only if a, b, p, d lie in the domain

154

SICE JCMSI, Vol. 1, No. 2, March 2008

D = {(a, b, p, d) | dα − dβ − p > 0}. When the system parameters lie in the domain D,

Jian-Xin XU

y(t p ) − A Mp = A ⎧  α−β ⎪ dα − dβ − p 2β 1 ⎪ ⎨ = (α + β)(dα − dβ − p) ⎪ ⎩ dα + dβ − p 2βp ⎪ ⎛ ⎫ α+β ⎪ ⎜⎜⎜ dα − dβ − p  2β ⎪ ⎬ ⎜ − ⎜⎜⎝ (α − β)(dα + dβ − p)⎪ . ⎪ ⎭ dα + dβ − p Now derive the settling time. Let | y(t)−A A | < 0.02 for all t ≥ t s . If the overshoot exceeds 2%, then t s is the bigger real root of the following equation     1 d 1 d p p −(α+β)t − + − e − e−(α−β)t 2 β β(α + β) 2 β β(α − β) p = 0.02 ; b+ p otherwise, t s is the unique real root of the following equation     p p 1 d 1 d −(α+β)t − − − + e e−(α−β)t 2 β β(α + β) 2 β β(α − β) p . = −0.02 b+ p Note that in neither cases can we obtain the analytic expression of t s . Critical Damped Case When the system parameters satisfy (a+kkd )2 −4(b+kk p ) = 0, the closed-loop system shows critical damped response with two identical real poles. The system response in the time domain is y(t) = A −

p + (p − dα)αt −αt e b+ p

(A. 5)

with the steady-state A. First derive the peak overshoot. Letting dy α2 (d + (p − dα)t) −αt = e =0 dt b+ p yields a unique solution tp =

−2d −d = , p − dα 2p − ad − d2

implying that the overshoot exists if and only if a2 < 4b + d2 . The overshoot is Mp =

d2 (a + d)2 − p(−a2 + 4b + 3d2 + 2ad) −2d(a+d) e −a2 +4b+d2 . p(−a2 + 4b + d2 )

Next derive the settling time. If the overshoot exceeds 2%, then t s is the bigger real root of the following equation −

p + (p − dα)αt −αt p e = 0.02 ; b+ p b+ p

otherwise, t s is the unique real root of the following equation −

p + (p − dα)αt −αt p e = −0.02 . b+ p b+ p

Note that in neither cases can we obtain the analytic expression of t s .

Received the Bachelor degree from Zhejiang University, China in 1982. He attended the University of Tokyo, Japan, where he received the Master’s and PhD degrees in 1986 and 1989 respectively. All degrees are in Electrical Engineering. He worked for one year in the Hitachi research Laboratory, Japan; for more than one year in Ohio State University, U.S.A. as a visiting scholar; and for 6 months in Yale University as a visiting research Fellow. In 1991 he joined the National University of Singapore, and is currently a professor in the Department of Electrical and Computer Engineering. He is a senior member of IEEE.

Deqing HUANG Received the Bachelor and PhD degrees from Mathematical College Sichuan University, China in 2002 and 2007 respectively. Currently he is a PhD student at the Department of Electrical and Computer Engineering, National University of Singapore. His research interests lie in the area of iterative learning control.

Srinivas PINDI Has done his BE in Electrical and Electronics from Osmania University, Hyderabad, India. after he worked as Scientific Officer in Bhabha Atomic Research Centre (BARC), India for 2 years. Currently pursuing his Master of Technology from Indian Institute of Technology Guwahati.

Suggest Documents