Robust digital design of continuous-time nonlinear control systems using adaptive prediction and random-local-optimal NARMAX model Zhi-Ren Tsai Department of Computer Science & Information Engineering, Asia University, Taiwan Graduate Institute of Biostatistics, China Medical University, Taiwan ren@asia.edu.tw Abstract In this paper the time-delay and uncertainty of continuous-time (CT) systems are considered, and it is suggested that input and output of a discrete-time (DT) Neural Plant Model (NPM) and recursive neural controller have scaling factors which limit the value zone of measured data from a system. Adapted scaling factors cause the tuned parameters to converge to obtain a robust control performance. However, the proposed Random-LocalOptimization (RLO) design for a model/controller uses off-line initialization to obtain a near global optimal model/controller. Other important issues are the considerations of cost, greater flexibility, and highly reliable digital products for these control problems. This issue of DT control design for CT plant is more difficult than that of CT control design for CT plant, because of the need to process the modeling error between the CT plant and DT model. The input-delay, uncertainty, and sampling distortion of a CT nonlinear power system need to be solved by developing a digital model-based controller. Here, this is called the DT tracking control design of CT systems (DT-CT). Therefore, the DT structure of the adaptive controller for the CT nonlinear power system should be designed as a kind of feed-forward-Recursive-Predictive controller (FRP). First, due to the problem of delays, a digital neural controller with feed-forward of the 1 reference signal and a Nonlinear Auto-Regressive Moving Average eXogenous (NARMAX) neural model design is adopted to reduce this difficulty. The most important contribution is that the more reasonable and systematic two-stage control design, the CT nonlinear delayed system to be controlled is modeled using a NARMAX technique with the first-stage (off-line) method by the proposed global optimal network algorithm and second-stage (online) adaptive steps. Second, the dynamic response of the system is controlled by an adaptive NARMAX neural controller via a sensitivity function. A theorizing method is then proposed to replace the sensitivity calculation, which reduces the calculation of Jacobin matrices of the BP method. Finally, the feed-forward input of reference signals helps the digital neural controller to improve the control performance, and the technique works to control the CT systems precisely. Keywords: Random-local-optimization algorithm, controller, DT-CT design 2 NARMAX model-based neural 1. Introduction During the past decade optimal control [1, 2] has attracted great attention from both the academic and industrial communities, and there have been many successful applications. Despite this success, it has become evident that many basic and important issues [3] remain to be further addressed. Of these, stability analysis and systematic designs are among the most important issues for optimal control systems [4] and robust control theories [1, 5-8], and there has been significant research on these issues (see [4, 9, 10]). In addition, a neural controller has been suggested as an alternative approach to conventional PID control techniques [11] for complex control systems [1]. Moreover, Neural-Network (NN) based modeling has become an active research field because of its unique merits in solving complex nonlinear system identification and control problems [10]. Neural networks (NNs) or NARMAX neural networks [12] are composed of simple elements operating in parallel, inspired by biological nervous systems. A neural network can be trained to represent a particular function by adjusting the weights between elements. Due to discrete-time (DT) controllers being cheaper and more flexible than continuous-time (CT) controllers, the DT control problem for CT plant is worth studying. In modern control engineering, controllers are commonly implemented directly by the hardware or software of digital computers. However, one important issue has to be faced; that is, the new design (DT-CT design) problem effects a new type of application, and an adaptive NN-model-based design method has not yet been developed to adjust the parameters of a discrete-time (DT) adaptive neural controller such that the original continuous-time (CT) system, with time delays and uncertainties, is uniformly ultimately bounded (UUB) stable. The study of CT control of CT time-delay systems has received considerable attention in recent years since delay is a major cause of poor performance in many important 3 engineering systems [13-15]. Hence, the future direction of CT time-delay control systems needs to involve the DT control problem. The amount of delay has different impacts on the various approaches [15-19]. As is known, the delay control problem is an important and complex factor in the stability performance of CT nonlinear systems. In general, a delay signal happens in a signal’s long-distance or heat translation. Based on the timer of the micro-controller or Digital Signal Process (DSP) chip, the effect of delay in neural system identification can be approximated by many tape-delay terms. This reduces the difficulty of delay identification. The DT NARMAX model is general sufficient to approximate an unknown, nonlinear, dynamical and delayed CT system by selecting an appropriate sampling time. DT control design for DT plant [20] and CT control design for CT plant are two kinds of well-known problems. [20] has inspired consideration of the more difficult problem of DT control design for CT plant, because of the need to process the modeling error between the CT plant and DT model, except for proposing the novel adaptive control law. The modeling and controlling performance can be guaranteed by the appropriate sampler and some theories. The modeling performance of plant is important for this research. The stability of results and the robust control design are related to this precise plant model. Hence, a two-stage training scheme is needed to guarantee a well-behaved model, referred to as a predictive controller. Based on this correct model, the control parameters can be updated by the BP method. That is one contribution of this paper. Another contribution is the proposal of a theorizing method to replace the sensitivity calculation or reduce the calculation of Jacobin matrices of the BP method. That is why the adaptive prediction control method is used to improve the DT control performance of the proposed DT-CT design by tuning the parameters of the model and controller. 4 The feed-forward term in [20] is derived indirectly by assuming many constraints, and due to the over-fitting and local optimal problems of NN modeling, the method [20] is not suitable for on-line applications because of the need for a lengthy convergence time. Therefore, to satisfy the on-line working requirements for accurate modeling of the plant, the NARMAX plant and control models are trained by initially using off-line methods. On the other hand, these neural techniques [21-23] have usually been demonstrated under nonlinear control due to their powerful nonlinear modeling capability [24] and adaptability. However, they must exhibit the optimal problem of falling into the local minimum easily by using the Back-Propagation (BP) or Levenberg-Marquardt BP algorithm (LMBP) [25] method. Hence, the RLO algorithm is proposed to improve this drawback. It not only guarantees the gradient decent method [26] against the local optimal solution, but also speeds up the convergence of the Particle Swarm Optimization (PSO) [31, 32]. Inspired by the DT neural controller of [20] only for a DT system, a digital neural control design for a CT system is proposed and an approximate inverse of the delayed plant dynamics is used to act as the NARMAX neural controller. The adaptive controller and NARMAX models are easier to converge than [21-23] by the proposed two-stage scheme. Moreover, the modeling error between the model and physical system is considered in the theorems by Lyapunov functions [27]. This paper concludes with a simulation example and experimental data to demonstrate these techniques. 2. System description First, consider a general nonlinear system with delays; described as follows: P: x (t ) fCT ( x, u, t , t , ) and y (t ) g ( x) , where P shown in Fig. 1 is a controlled plant; the bounded uncertainties 5 (1) (t ) create the dynamic quality of the system parameters which refer to electrical elements of the power system; the control input u (t ) ; is the time delay; g () is the relational function of the state x(t ) and system output y (t ) . Then, f CT is discretized by setting the appropriate sampling time or sampling period Ts (sec) of DTC-CTP design, and t k Ts , where k is a positive integer, to the following DT nonlinear system f DT : x(( k 1) Ts ) f D T ( x(k Ts ), u (k Ts ), Ts ), y(k Ts ) g ( x(k Ts )) or x(k 1) f D T ( x(k ), u (k ), Ts ), y (k ) g ( x(k )) , (2) where k indicates the signal sequence, the DT state vector is x(k Ts ) , and the DT control input is u (k Ts ) . The zero-order-hold control input u (t ) u (k Ts ) u (k ) , where k is also the index of the discrete result u (k ) of u (t ) referring to the NN model P̂ (see Fig.1) of (1). In this control structure in Fig. 1, the NN plant model is designed to approximate this nonlinear system Eq.(1). This NN plant model, or control model is built following the subsequent mathematical equations. An NN plant model or control model with L layers each having N l ( r 1l ,2l ,..., N l ; l 1,2,..., L ) neurons is established to approximate a DT nonlinear system f DT Eq.(2). Superscripts are used to distinguish between these layers. Specifically, the number of the layer is appended as a superscript to the name of the variable. Thus, the weight matrix for the l -th ( l 1,2,..., L ) layer is written as W l and {WPl ,WCl } W l . Moreover, it is assumed that vr ( r 1l ,2l ,..., N l ) is the net input and that all the transfer functions Tr (vr ) of units in the NN system are described by the following function: 2 Tr (vr ) 1, 1 exp( vr / q) for 6 r 1l ,2l ,..., N l and l L ; Tr (vr ) vr , where q and r 1L ,2 L ,..., N L ; for are positive parameters associated with the sigmoid function. The transfer function vector of the l -th layer is defined as: l (vr ) [Tr (v1 ), Tr (v2 ),..., Tr (vN )]T , l 1,2,..., L , l l l where Tr (vr ) ( r 1l ,2l ,..., N l ) is a transfer function of the r -th neuron. The final outputs of the NN plant model P̂ and control model C F can then be inferred as follows: respectively: yˆ (k ) L (W L L 1 (W L 1 L 2 (... 2 (W 2 1 (W 1Z (Ts )))...))) Pˆ ( y(k 1), y(k 2),..., y(k n), u(k ), u(k 1),..., u(k p), WP (k ), Ts ) , and u F C F (u (k 1), u (k 2),..., u (k cu ), r (k ), r (k 1),..., r (k c y ), WC (k ), Ts ) , where r (k ) is a reference input, Z T (Ts ) [ g ( x(( k 1) Ts )), g ( x(( k 2) Ts )),..., u (k Ts ), u (( k 1) Ts ),...,1] , the adaptive parameter WP (k ) [WP1 ,WP2 ,...] or WC (k ) [WC1 ,WC2 ,...] of neural weights’ and biases’ refers to the iteration k and the proposed adaptive laws for the controller and plant models are as follows: WC (k 1) WC (k ) WC (k ) , and WP (k 1) WP (k ) WP (k ) . Although WP (k ) and WC (k ) are the proposed adaptive laws of plant model and control model, respectively, where: W P (k ) P ( yˆ (k ) y (k )) but implementing dyˆ (k ) , dW P (k ) WC (k ) T C ( yˆ (k ) r (k )) dyˆ (k ) , dWC (k ) dyˆ (k ) needs too many Jacobin matrices’ calculations, so the following dWC (k ) T adaptive prediction control law is used WC (k ) X u X 7 du (k ) to replace the above dWC (k ) adaptive laws to reduce computing time, where P , C , X are learning rates; u X C X (e(k )) [ K 1 (e1 (k )), K 2 (e2 (k )), K 3 (e3 (k )),...] T is the predictor output, where the tracking error is e(k ) r (k ) y (k ) , and e(k ) [e1 (k ), e2 (k ), e3 (k ),...]T , K1 (), K 2 (), K 3 (),... are defined by the user. A composite controller, u(k ) u F s u X , where s {0,1} is proposed in the next section. 3. Control architecture, neural-model-based controller design and control scheme 3.1 Adaptive digital neural controller design through neural plant model In this paper an adaptive prediction control structure is proposed, as shown in Fig. 1, where the FRP controller CF is designed as follows: u(k ) u( z (k )) C F ( z (k ), WC (k ), Ts ) s C X (e(k )) u F s u X , 0, where e(k ) r (k ) y (k ) , the switch index s 1, ~ e (k 1) 0, And, ~ e (k 1) 0. u( zˆ(k )) C F ( zˆ(k ), WC (k ), Ts ), where uP uF u X , (3) yˆ 1 (k 1) Pˆ (u P (k 1)) eˆ1 (k 1) r (k 1) yˆ1 (k 1) , eˆ2 (k 1) r (k 1) yˆ 2 (k 1) , (4) , yˆ 2 (k 1) Pˆ (u F (k 1)) , eˆ1 (k 1) eˆ2 (k 1) ~ e (k 1) , as shown in Fig. 1. The feed-forward terms are reference signals [r (k ), r (k 1),..., r (k p)] , and recursive terms are control signals [u (k 1), u (k 2),..., u (k q)] . The off-line training input of controller is: zˆ(k ) [ y (k ), y (k 1),..., y (k p), u (k 1), u (k 2),..., u (k q)] , The on-line recursive input of controller is: z (k ) [r (k ), r (k 1),..., r (k p), u (k 1), u (k 2),..., u (k q)] . The controller has two working phases: z (k ) is the data vector of the testing phase, and zˆ ( k ) is the data vector of the training 8 phase. The tuned parameter vector of the controller is: WC (k ) of (3). The proposed on-line digital neural controller [r (k ), r (k 1),..., r (k p)] and recursive structure u (k ) has feed-forward terms [u (k 1), u (k 2),..., u (k q)] . Hence, it uses a NARMAX neural model or inverse of the plant dynamics to aid control precision in the face of a delayed plant with uncertainties. Adapting the neural controller can suppress the uncertainty of the plant P shown in Fig. 1. Although the structure of the neural controller is chosen as (3), the neural controller has not been designed because the parameter vector WC (k ) is not specified. Ts is the chosen tape-delay time, is a positive integer. The idea of the inverse-model-based neural controller is proposed by the following simplified relation: If y (k ) Pˆ (u (k )) , u (k ) Pˆ 1 (r (k )) C F (r (k )) , then y (k ) r (k ) , (5) where Pˆ () is the adaptive NARMAX neural model of plant; CF () is the adaptive NARMAX neural controller; r (k ) is the desired output. According to the idea of Eq.(5), the recursive structure Pˆ () can be designed with tape delays as follows: y(k ) yˆ (k ) Pˆ ( yˆ (k 1), yˆ (k 2),..., yˆ (k n), u(k ), u(k 1),..., u(k p),WP (k ), Ts ) , (6) ˆ u , respectively. where n, p 1 are the amount of tape delays of y, But, due to the parameters of the recursive structure are converged much harder, the weights and biases WP (k ) of this model are trained by the feed-forward structure as follows: yˆ (k ) Pˆ ( y(k 1), y(k 2),..., y(k n), u(k ), u(k 1),..., u(k p),WP (k ), Ts ) . (7) The plant output is compared with the desired output to create a tracking error signal e(k ) r (k ) y (k ) . The system errors eˆ(k ) r (k ) yˆ (k ) and e(k ) are used by the adaptation algorithm to update the parameters of P̂ and CF . Next, the performance index 9 for minimizing the tracking error is designed, as follows: 1 1 1 J (k ) e(k )T e(k ) (r (k ) y (k )) T (r (k ) y (k )) ( y (k ) r (k )) T ( y (k ) r (k )) , 2 2 2 (8) is a simple cost function to be minimized by the proposed algorithm. Then, the on-line BP algorithm adapts the control parameter matrix WC (k ) . That is, the change in control parameters WC (k ) is calculated as dJ (k ) d ( y (k ) r (k ))T ( y (k ) r (k )) WC (k ) C (k ) C (k ) dWC (k ) 2 dWC (k ) T C (k )( y (k ) r (k )) dy (k ) dy (k ) du (k ) , C (k )( y (k ) r (k )) dWC (k ) du (k ) dWC (k ) (9) where the small positive C (k ) can be selected as a stable learning rate via the following theorems. Theorem 1: If the number of neurons and tape-delay terms of the neural model is sufficient, and the appropriate sampling time Ts is selected to let y (k ) y(k ) and the following condition 0 P (k ) 2 dyˆ (k ) dWP (k ) 2 P , (10) is satisfied, where py pu dyˆ (k ) yˆ (k ) yˆ (k ) du (k i ) yˆ (k ) dyˆ (k i ) ; dWP (k ) WP (k ) i 0 u (k i ) dWP (k ) i 1 yˆ (k i ) dWP (k ) y (k ) is the output of the optimal model, then the trajectories yˆ (k ) converging to plant output y (k ) is a uniformly ultimately bounded (UUB) approximation on the bounded error yˆ (k ) y (k ) . 10 3.2 Proof of Theorem 1 First, consider the following ideal Lyapunov candidate [27] for the model part, 1 1 V1 (k ) ( yˆ (k ) y (k )) T ( yˆ (k ) y (k )) yˆ (k ) y (k ) y (k ) y (k ) 2 2 where V2 (k ) 1 yˆ (k ) y (k ) 2 2 2 1 2 yˆ (k ) y (k ) (k ) V2 (k ) (k ) , 2 (11) is an actual Lyapunov candidate of reachable and assumptive trajectory y (k ) , the bounded approximation error (k ) 1 2 y (k ) y (k ) ( yˆ (k ) y (k )) T ( y (k ) y (k )) , 2 and the number of neurons of the neural model is sufficient and the appropriate sampling time Ts is selected to let y (k ) y (k ) . The next task is to train this neural model such that V2 (k ) is minimized, WP (k ) dyˆ (k ) ( yˆ (k ) y (k )) P (k ) dWP (k ) dV2 (k ) d ( yˆ (k ) y (k )) dyˆ (k ) ( yˆ (k ) y (k )) ( yˆ (k ) y (k )) . dWP (k ) dWP (k ) dWP (k ) (12) Then, the following Lyapunov candidate for the controller is designed: V3 (k ) 1 1 2 ( yˆ (k ) r (k ))T ( yˆ (k ) r (k )) yˆ (k ) r (k ) , 2 2 (13) thus the change in the Lyapunov function is obtained by: V3 (k 1) V3 (k ) 1 2 2 ( yˆ (k 1) r (k 1) yˆ (k ) r (k ) ) . 2 (14) Finally, the update law of the control parameters of the controller is obtained as follows: WC (k )T dV (k ) dyˆ (k ) 3 ( yˆ (k ) r (k )) . C (k ) dWC (k ) dWC (k ) 11 (15) This study develops some convergence theorems to select appropriate stable learning rates. First, the difference of modeling error eP (k ) yˆ (k ) y(k ) can be represented by T de (k ) T deP (k ) deP (k ) de (k ) eP (k 1) eP (k ) P (k )eP (k ) eP (k )1 P P (k ) P dWP (k ) dWP (k ) dWP (k ) dWP (k ) dyˆ (k ) T dyˆ (k ) eP (k )1 ( k ) , dWP (k ) P dWP (k ) (16) thus the change in the Lyapunov function is obtained by: 2 dyˆ (k ) T ˆ 1 1 d y ( k ) 2 2 2 V2 (k 1) V2 (k ) ( eP (k 1) eP (k ) ) eP (k ) 1 (k ) eP (k ) dWP (k ) P dWP (k ) 2 2 2 T ˆ ˆ 1 d y ( k ) d y ( k ) 2 eP ( k ) 1 (k ) 1 . dWP (k ) P 2 dWP (k ) Hence, if dyˆ (k ) T dyˆ (k ) 1 1 P (k ) 1 dWP (k ) dWP (k ) and y (k ) y (k ) , then V2 (k 1) V2 (k ) , that is V2 (k ) 0 or yˆ (k ) y(k ) , makes the UUB approximation of this model on the bounded yˆ (k ) y (k ) . The proof is thereby completed. Furthermore, the following theorem for the convergence of the controller is obtained by the same procedure as the above proof. Theorem 2: If Theorem 1 in Eq.(10) is satisfied, the function to let the following condition, 12 dyˆ (k ) in Eq.(15) is computed dWC (k ) 0 C (k ) 2 dyˆ (k ) dWC (k ) 2 C , (17) be satisfied. Where with py pu dyˆ (k ) yˆ (k ) du (k i ) yˆ (k ) dyˆ (k i ) , dWC (k ) i 0 u (k i ) dWC (k ) i 1 yˆ (k i ) dWC (k ) cu du (k ) u (k ) u (k ) du (k i) , dWC (k ) WC (k ) i1 u (k i ) dWC (k ) then the nonlinear systems (1) in Fig. 1c are UUB stable, and the tracking errors e(k ) r (k ) y (k ) are bounded via the controller. Hence, the dynamic response of the system P can be controlled using CF , as shown in Fig. 1. This CF needs the plant model P̂ to adjust control parameters via sensitivity function yˆ (k ) . u (k i ) The digital feedback controller includes a delay block D, as shown in Fig. 1. Here, the error e~ (k 1) is used to estimate u X , and the proposed predictor of the delayed system can let us cancel some complex computations, such as yˆ (k 1) yˆ (k 1) yˆ (k ) yˆ (k 1) yˆ (k 1) , u (k 1) u (k 1) u (k 1) uX of sensitivity function yˆ (k ) in the BP algorithm. Hence, the following theorem is u (k i ) proposed to update the control parameters of FRP under the assumption of providing a model which applies a lower prediction error, and a more correct eˆ(k 1) r (k 1) yˆ (k 1) u X . The prediction error is bounded, due to the previous eˆ(k ) r (k ) yˆ (k ) being bounded at any time. Hence, the prediction error eˆ( k 1) will be bounded by using Theorem 1-2. Furthermore, the following theorem is obtained for the convergence of the 13 adaptive prediction controller by the same procedure as Theorem 1. Theorem 3: If Theorem 1 in Eq.(10) is satisfied, the predictive function du (k 1) is dWC (k 1) computed to let the following condition, 0 X (k 1) 2 du (k 1) dWC (k 1) 2 X , (18) be satisfied, then the nonlinear systems (1) in Fig. 1a-1b are UUB stable, and the tracking errors e( k ) r ( k ) y ( k ) are bounded via the predictive controller u(k 1) uF (k 1) s u X . The tracking error is e(k ) r (k ) y (k ) , and e(k ) [e1 (k ), e2 (k ), e3 (k ),...]T , but the parameters of adaptive control C F are updated by using the predictive offset u(k 1) u p u(k ) uX CX (e(k )) [ K1 (e1 (k )), K2 (e2 (k )), K3 (e3 (k )),...]T , of the prediction control input u p u (k 1) , where replaces yˆ (k 1) yˆ (k 1) yˆ (k ) yˆ (k 1) u(k 1) u p u (k ) uX dyˆ (k ) in the following recursive equation: du (k ) py pu dyˆ (k ) dyˆ (k ) du (k ) yˆ (k ) du (k i ) yˆ (k ) dyˆ (k i ) , dWC (k ) du (k ) dWC (k ) i 0 u (k i ) dWC (k ) i 1 yˆ (k i ) dWC (k ) du ( k ) or dWC ( k ) therefore, only du ( k 1) need to be calculated to update WC (k ) or WC (k 1) , respectively, dWC ( k 1) and K1 (), K 2 (), K 3 (),... are defined by the user. 3.3. Two-stage scheme Fig. 1 shows a block-diagram of an adaptive recursive control system. The system to be 14 controlled is labeled as the plant P , which is subject to uncertainties and delays. Due to gradient-descent based training algorithms, let the model/controller converge to a local minimum in the solution space. Hence, the two-stage training algorithm is proposed, as follows. In the first stage, the measured data is used to train the global optimal NARMAX plant and neural controller by the training-data-shuffle method. This method shuffles the training data to avoid most of the local optimal solutions obtained by the off-line training procedure in next section. The measured data is divided into a training data and other testing data. This testing data is not used for training the NN. However, the final performance of the NN is decided by the testing data and the training data. In the second stage, the global optimal NARMAX plant model and neural controller is adapted. The two stages are divided into the following five steps: Step 1: First, the reference signal, r (k ) , is designed. By the white noise of input u (k ) for plant, output data y (k ) is collected and a training-data-shuffle method is used to shuffle the input/output pairs’ data. These shuffled data are ready to train the NARMAX model/controller. Here, the following reasonable conditions need to be taken into account: max (r (k )) max ( y (k )) , min (r (k )) min ( y (k )) , max (u (k )) uU , and min (u (k )) u L ,(19) k k k k k k need be satisfied, where uU is the upper bound of u (k ) , and u L is the lower bound of u (k ) . According to Eq.(19), much of the excessive control effort u (k ) can be avoided. If Eq.(19) is satisfied, then go to Step 2. Step 2: The feed-forward structure model P̂ is trained/tested off line yˆ (k ) Pˆ (Su u(k ), Su u (k 1),..., Su u (k pu ), S y y(k 1), S y y(k 2),..., S y y(k p y ), WP (k ), Ts )(1 / S y ), 15 (20) via the shuffled input/output pairs’ data. After system identification P̂ is performed, and the digital neural controller CF for the CT system can be built by using this inverse NARMAX plant model Pˆ 1 in the next step. Step 3: In practice, according to the exchanged output/input pairs’ data from Step 2, the offline stage to train/test the neural controller can be passed through u (k ) CF ( Su u (k 1), Su u (k 2),..., Su u (k cu ), S y y (k ), S y y (k 1),..., S y y (k c y ), WC (k ), Ts )(1 / Su ) . (21) If Eq.(20) and Eq.(21) work, go to Step 4. Step 4: Update the on-line weights and biases WP of the recursive structure model P̂ : yˆ (k ) Pˆ (Su u(k ), Su u (k 1),..., Su u (k pu ), S y yˆ (k 1), S y yˆ (k 2),..., S y yˆ (k p y ), WP (k ), Ts )(1 / S y ), (22) to approximate the CT nonlinear system by using Remark 1 and Theorem 1. Due to the adaption laws for Eq.(20) and Eq.(22), an exchange for both of them can be designed to switch into the system, as a switching in Fig. 1, when Eq.(22)’s absolute approximation error is too big. If Eq.(20) and Eq.(22) work, then go to Step 5. Step 5: Adapt the digital neural controller for the modeling error and tracking error by using Remark 1 and Theorem 1-2. Finally, update the on-line parameters of the neural controller C F u (k ) CF ( Su u (k 1), Su u (k 2),..., Su u (k cu ), S y r (k ), S y r (k 1),..., S y r (k c y ), WC (k ), Ts )(1 / Su ) , (23) to minimize the tracking error, and finish the above two stages: the off-line stage and on-line stage. 16 To make sure of the robustness of the control system, the convergence to the global optimal solution of parameters of the model/controller has to be guaranteed. Hence, some random initial weights and biases of the model are designed by Particle Swarm Optimization (PSO) [31, 32] with the parameters of the controller first. The PSO algorithm consists of the velocity vi ( j 1) vi ( j ) 1i ( pi xi ( j )) 2i (G xi ( j )) , and position xi ( j 1) xi ( j ) vi ( j 1) , where i 1,2,..., H is the particle index; j 1,2,..., N is the iteration index; vi is the velocity of ith particle; xi is the position of ith particle; pi is the best position found by ith particle (personal best); G is the best position found by the swarm (global best, best of personal best); 1i , 2i are the random numbers on the interval [0,1] applied to the ith particle. The PSO supplies random initial parameters, hence, it is an initial parameters’ conductor. These initial parameters are then converged locally by the LMBP method and the best solution for the initial model/controller is chosen. Finally, the global optimal solution of parameters can be found every time. Hence, this idea has been named the Random-Local-Optimization (RLO) algorithm. The RLO algorithm is a composite of the LMBP algorithm and a random initialization procedure of evaluating fitness value 1 /( 0.01) , where 1 (1 ) 2 , [0,1] . The total of absolute training error 1 is obtained by LMBP via the training data, and 2 is the total of absolute testing error of the model/controller output via the testing data input. In this paper, off-line RLO is used as a learning algorithm for the feed-forward structure model Eq.(20) due to the on-line tuning parameters of the recursive structure of the plant model being not converged. After the offline training stage, in order to tune on-line the parameters of the plant model Eq.(22) 17 recursively, dyˆ (k ) of Eq.(10) needs to be calculated as follows: dWP (k ) py pu dyˆ (k ) yˆ (k ) yˆ (k ) du (k i ) yˆ (k ) dyˆ (k i ) . dWP (k ) WP (k ) i 0 u (k i ) dWP (k ) i 1 yˆ (k i ) dWP (k ) (24) Similarly, in order to tune the on-line parameters of the controller Eq.(23) recursively, and dyˆ (k ) of Eq.(18) needs to be calculated as follows: dWC (k ) py pu dyˆ (k ) yˆ (k ) du (k i ) yˆ (k ) dyˆ (k i ) , dWC (k ) i 0 u (k i ) dWC (k ) i 1 yˆ (k i ) dWC (k ) where cu du (k ) u (k ) u (k ) du (k i) . dWC (k ) WC (k ) i1 u (k i ) dWC (k ) (25) (26) Hence, the following algorithm adapts a NARMAX neural controller for a NARMAX neural model of plant. Step 1: Back propagate through C F to form u (k ) u (k ) and in Eq.(26). If update WC (k ) u (k i ) du (k ) du ( k i ) of Eq.(26), and shift down in Eq.(25), then go to Step 2. dWC ( k ) dWC (k ) Step 2: Back propagate through P̂ to form update yˆ (k ) yˆ (k ) and in Eq.(25). If yˆ (k i ) u (k i ) dyˆ (k ) dyˆ ( k i ) of Eq.(25), and shift down in Eq.(25), then go to Step 3. dWC (k ) dWC ( k ) T Step 3: Compute WC (k ) C (k )( yˆ (k ) r (k )) dyˆ (k ) . If update weights dWC (k ) WC (k 1) WC (k ) WC (k ) , and WP (k 1) WP (k ) WP (k ) , then go to Step 1. To clarify this method, in [20], a robust and adaptive method was used to allow learning to occur on-line, tuning performance as the system runs. But, [20] didn’t consider the prediction, modeling error, global optimal initialization of control parameters, the problem 18 of lengthy convergence time of on-line control, delayed terms, uncertainties in plant and DT-CT problems. Moreover, the choice method of initial parameters of the on-line controller still lacks the ability to overcome the over-fitting problem of the controller. Hence, the offline stage is proposed for a RLO learning algorithm to choose the initial weights and biases of the on-line neural controller in the simulation example of the power plant, as shown in the following case study. 4. Cases study First, the conventional PWM buck converter, by using AM-OTS-DS [28, 29] methodology, is modeled to the following equivalent circuit plant: 1 R RC RL iL (t ) L R RC R vC (t ) C (R R ) C R 1 VD L ( R RC ) iL (t ) RM iL (t ) Vin VD L , u ( t ) L vC (t ) 1 0 0 C ( R RC ) R RC vo (t ) R RC R iL (t ) , R RC vC (t ) (27a) In this paper the robustness of the above control system is emphasized, so uncertainty , and delay are added to the original control system. 1 R RC R iL (t ) L L R RC R v ( t ) C C (R R ) C R 1 VD L ( R RC ) iL (t ) RM iL (t ) Vin VD L , u ( t ) L 1 vC (t ) 0 0 C ( R RC ) R RC vo (t ) R RC where R R ; R 6 R iL (t ) vo (t ), R RC vC (t ) ; Vin 30 is a DC voltage source; 19 (27b) C 202.5 10 6 ; 0.52 sin( 2 t / 3) ; 0.1 ; 102 (1 sin( 10t )) ; L 98.58 106 ; RL 48.5 103 and RC 162 103 are the parasitic resistances of the inductor and capacitor, respectively. The element RM 0.27 is the static drain to source resistance of the power MOSFET and VD 0.82 is the forward voltage of the power diode. u (t ) is the duty ratio of conventional PWM buck converter. x(t ) x [iL (t ), vC (t )]T is the state of the system, and the output of this power system is y(t ) vo (t ) . The nonlinear, uncertain, hotter circuit’s components, time-delay, and digital control problems of PWM buck converter CT system renders a tracking control problem difficult to analyze. A simulation system in Eq.(27) is built with uncertainty. In this study, it is assumed that the parameters of the circuit’s components are not ideal, and the capacity of the digital controller is limited by using a lower-cost chip. Here, the sampling period Ts 104 is designed for this power system Eq.(27b). Hence, the delay is very large for this system. Referring to Fig. 1, and the above sections, it can be seen how to model the plant dynamics by considering the modeling error, and how to use the neural model of plant to adapt a neural controller. To compare with other methods, the following cases are introduced: Case 1: This case is in [29], and its digital controller is a kind of T-S fuzzy controller with integral term. Case 2: This case is in [31], and its controller is a kind of PID, with PSO to compare with Case 3. Case 3: This is the control method presented here, and the proposed neural-modelbased neural controller is adaptive and globally optimal. The detail designs of the Case 1−Case 3 are as follows: 20 Case 1 is a LMI control method of original example for this power plant. The control parameters of Case 1 are as follows: K11 K 21 [0.0476, 0.9348]T , K12 K22 -426.4969 , and this T-S fuzzy controller of Case 1 is designed as: u (k ) h1 ( K11x(k ) K12 e(k )Ts ) h2 ( K 21x(k ) K 22 e(k )Ts ) . k k Case 2 is an optimal control method. This PID controller is designed as: u (k ) u (k 1) K p (e(k ) e(k 1)) K i Ts 1 (e(k ) e(k 1)) K d (e(k ) 2e(k 1) e(k 2)) . 2 Ts And, the parameters H 30, N 50 of PSO of Case 2. Case 3 is also an optimal control method, but its NARMAX neural control design method is very different from PSO of Case 2. The predictive controller of Case 3 is u X K1 (e(k )) , where K1 0.1 . T First, the initial state is set to x(0) [0,0] , and the reference signal r (k ) 12 sin( 30k Ts ) . Case 3 uses NN structure 5-8-1 of the NARMAX plant model, it has 5 inputs, [u (k ), u (k 1), u (k 3), yˆ (k 1), yˆ (k 2)] , 8 tansig(·) neurons in the hidden layer, 1 purelin(·) neuron in the output yˆ (k ) layer. Also, Case 3 uses NN structure 5-8-1 of the NARMAX controller, it has 5 inputs, [r (k ), r (k 1), r (k 3), u (k 1), u (k 2)] , 6 tansig(·) neurons in the hidden layer, and 1 purelin(·) neuron in the output u (k ) layer. The weights and biases, WP , are trained as follows by selecting two suitable scaling factors Su 1 and S y 1 of the plant model, whose summation WPS of WP 24 updated as shown in Fig. 2a. The summation WCS is of the neural control parameters WC is updated as shown in Fig.2a. The tracking control performance of Case 3 is shown in Fig.2b. 21 The optimal parameters K p -0.003241, Ki 34.280925, Kd 3.4004 10-6 of the PID controller of Case 2 are obtained by using PSO. The tracking control performance of Case 2 is shown in Fig.3. The tracking control performance of Case 1 is shown in Fig.4. Finally, the control performances of Case 1−Case 3 are compared, as shown in Fig. 5. Fig. 2 shows the precise neural control performance of Case 3. Fig. 3 shows the digital PID with PSO control performance of Case 2. Fig. 4 shows the LMI control performance of Case 1. It is clear that the proposed two-stage scheme, Case 3, has excellent tracking performance when compared with Case 1 and Case 2. Conclusions The proposed two-stage adaptive prediction control converges very quickly, works highly effectively, and precisely. It works for nonlinear delayed plants with uncertainty. The recursive and feed-forward control scheme is partitioned into two stages that can be independently optimized. First, an off-line neural model of a continuous-time (CT) nonlinear power plant is made; second, a constrained off-line digital neural controller is generated; then, an adaptive plant model is made, and an adaptive NARMAX neural controller with predictor is generated; finally, all processes may continue concurrently, and robustness and DT-CT problems for a power plant are solved. Acknowledgment The author would like to thank the National Science Council for it support under contracts: NSC-97-2218-E-468-009, NSC-98-2221-E-468-022, NSC-99-2628-E-468-023, NSC100-2628-E-468-001, NSC-101-2221-E-468-024 and Asia University under contract 98-ASIA22 02, 100-asia-35 and 101-asia-29. References [1] H. Dong, Z. Wang, D.W.C. Ho, H. Gao, “Robust H∞ neural output-feedback control with multiple probabilistic delays and multiple missing measurements,” IEEE Trans. Neural Syst., vol. 18, pp. 712-725, 2010. [2] L.X. Wang, A Course in Neural Systems and Control, Prentice-Hall, New Jersey, 1997. [3] Z. Wang, Y. Liu, G. Wei, X. Liu, “A note on control of a class of discrete-time stochastic systems with distributed delays and nonlinear disturbances,” Automatica, vol. 46, pp. 543548, 2010. [4] H.O. Wang, K. Tanaka, M.F. Griffin, “An approach to neural control of nonlinear systems: stability and design issues,” IEEE Trans. Neural Syst., vol. 4, pp. 14-23, 1996. [6] B. Shen, Z. Wang, H. Shu, G. Wei, “Robust H∞ finite-horizon filtering with randomly occurred nonlinearities and quantization effects,” Automatica, vol. 46, pp. 1743-1751, 2010. [9] B.S. Chen, C.S. Tseng, H.J. Uang, “Mixed H2 /H∞ neural output feedback control design for nonlinear dynamic systems: an LMI approach,” IEEE Trans. Neural Syst., vol. 8, pp. 249-265, 2000. [10] K. Tanaka, “An approach to stability criteria of neural-network control systems,” IEEE Trans. Neural Networks, vol. 7, pp. 629-643, 1996. [12] H.T. Siegelmann, B.B. Horne, C.L. Giles, “Computational capabilities of recursive NARX neural networks,” IEEE Trans. Syst., Man, Cybern. B, vol.27, pp.208-215, 1997. [13] K.R. Lee, J.H. Kim, E.T. Jeung, H.B. Park, “Output feedback robust H∞ control of uncertain neural dynamic systems with time-varying delay,” IEEE Trans. on Neural Syst., vol.8, pp.657-664, 2000. 23 [14] Y.Y. Cao, P.M. Frank, “Analysis and synthesis of nonlinear time-delay systems via neural control approach,” IEEE Trans. on Neural Syst., vol.8, pp.200-211, 2000. [15] C. Lin, Q.G. Wang, T.H. Lee, “Delay-dependent LMI conditions for stability and stabilization of T-S neural systems with bounded time-delay,” Neural Sets and Syst., vol.157, pp.1229-1247, 2006. [16] B. Chen, X. Liu, “Delay-dependent robust H∞ control for T-S neural systems with time delay,” IEEE Trans. on Neural Syst., vol.13, pp.544-556, 2005. [17] J. Yoneyama, “Design of H∞ control for neural time-delay systems,” Neural Sets and Syst., vol.151, pp.167-190, 2005. [18] S. Hu, Y. Liu, “Robust H∞ control of multiple time-delay uncertain nonlinear system using neural model and adaptive neural network,” Neural Sets and Syst., vol.146, pp.403-420, 2004. [19] H.J. Lee, J.B. Park, Y.H. Joo, “Robust control for uncertain Takagi-Sugeno neural systems with time-varying input delay,” Journal of Dynamic Syst., Measurement and Control, Trans. of the ASME, vol.127, pp.302-306, 2005. [20] G.L. Plett, “Adaptive inverse control of linear and nonlinear systems using dynamic neural networks,” IEEE Trans. on Neural Networks, vol.14, pp.360-376, 2003. [21] C.T. Chen, S.T. Peng, “Intelligent process control using neural techniques,” Journal of Process Control, vol.16, pp.493-503, 1999. [22] F.J. Lin, W.J. Hwang, R.J. Wai, “A neural network control system for tracking periodic inputs,” IEEE Trans. Neural Syst., vol.7, pp.41-52, 1999. [23] Y.C. Chen, C.C. Teng, “A model reference control structure using a neural network,” Neural Sets and Syst., vol.73, pp.291-312, 1995. [24] C. Li, K.H. Cheng, “Recursive neural hybrid-learning approach to accurate system 24 modeling,” Neural Sets and Syst., vol.158, pp.194-212, 2007. [25] M.T. Han, H.B. Demuth, M. Beale, Neural Network Design, PWS, 1996. [26] J.S.R. Jang, C.T. Sun, E. Mizutani, Neural and Soft Computing, Prentice-Hall, 1997. [27] J.P. LaSalle, “Some extensions of Lyapunov’s second method,” IRE Trans. Circuit Theory, pp. 520-527, 1960. [28] J. Sun and H. Grotstollen, “Averaged modeling of switching power converters: reformulation and theoretical basis,” IEEE Conf. Syst., pp. 1165-1172, 1992. [29] C.Y. Huang, “T-S neural controller design for DC-DC power converter,” Master Thesis, Chung Yuan Christian University, Taiwan, 2002. [30] Z.R. Tsai, Y.Z. Chang, J.D. Hwang, J. Lee, “Robust neural stabilization of dithered chaotic systems using IROA,” Information Sciences, vol. 178, pp.1171-1188, 2008. [31] W.D. Chang, “PID control for chaotic synchronization using particle swarm optimization,” Chaos, Solitons and Fractals, vol. 39, pp.910-917, 2009. [32] W.D. Chang, S.P. Shih, “PID controller design of nonlinear systems using an improved particle swarm optimization approach,” Commun Nonlinear Sci Numer Simulat, vol. 15, pp. 3632-3639, 2010. 25 Predictor RLO off-line FRP Plant Sw Sampler ZOH on-line off-line NPM Switching (a) Sampler NARMAX neural D NPM ZOH Plant |∙| NPM |∙| (b) 26 on-line RLO off-line FRP Plant Sampler ZOH on-line off-line NPM on-line Switching (c) Figure 1. (a) The proposed two-stage adaptive prediction structure of DT-CT control system. (b) On-line adaptive prediction block diagram for Theorem 1 and Theorem 3. (c) On-line adaptive block diagram for Theorem 1 and Theorem 2. 27 43.535 50.045 43.53 50.04 43.525 W PS W CS 50.05 50.035 43.52 50.03 43.515 43.51 50.025 0 0.05 0.1 0.15 0.2 0.25 0.3 0 0.35 0.05 0.1 0.15 0.2 0.25 0.3 0.35 t = k T (sec) t = k T (sec) s s (a) 20 r 10 k y 0 k -10 -20 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.1 rk - yk 0 -0.1 -0.2 -0.3 t = k T (sec) s (b) Figure 2. (a) The learning curve of the summation of WP , the learning curve of the summation of WC , and (b) the tracking control performance of Case 3. 28 20 r 10 k y 0 k -10 -20 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 1.5 rk - y k 1 0.5 0 -0.5 t = k T (sec) s Figure 3. The tracking control performance of Case 2. 29 15 r k 10 y k 5 0 -5 -10 -15 0 0.05 0.1 0.15 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.2 0.25 0.3 0.35 1 0.8 0.6 rk - yk 0.4 0.2 0 -0.2 -0.4 -0.6 -0.8 t = k T (sec) s Figure 4. The LMI control performance of Case 1. 30 1.2 Case 1 1 Case 2 Case 3 0.8 0.6 rk - yk 0.4 0.2 0 -0.2 -0.4 -0.6 -0.8 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 t = k T (sec) s Figure 5. Comparison of the control performances for Case 1 to Case 3. 31