2. College of Information Science and Technology, Beijing University of Chemical Technology, Beijing 100029, China
A kind of datadriven control method has been proposed recently^{[1]},i.e.,the model free adaptive control (MFAC) algorithm which is composed of the dynamic linearization data model (DLDM) and the "control function". Notably,the DLDM does not depend on the concrete mathematical model of the controlled object,which can be calculated based on the output$'$s pseudo gradient at each sample time.
The MFAC was proposed for unknown nonlinear systems,and has been successfully applied to many process control systems^{[2]}. For the MFAC approach,an equivalent dynamic linearization data model is built at each operation point of the closedloop system,and then the MFAC controller is designed based on this equivalent data model. Compared with traditional adaptive control methods,no structure information about the plant model is needed during the design of controller and estimator,and no unmodelled dynamics exists. Therefore,the MFAC is suitable for many practical systems since it is robust to system uncertainty.
As one of the datadriven control method,the MFAC has received much attention. For example,most of the MFAC control laws for singleinput single output (SISO) systems can be directly extended to complex multiinput multioutput (MIMO) systems^{[3]}. Inspired by the work of dynamic linearization technique of Hou et al.^{[2]},a multiple adaptive observerbased control strategy for nonlinear processes is proposed in which the pseudopartial derivative (PPD) theory is used to dynamically linearize the MIMO nonlinear system^{[4]}.
However,the MFAC is still developing^{[5]}. The proof of stability and convergence of the tracking problems and parameter tuning are its two open problems. In general,the parameters of MFAC are selected only according to qualitative analysis^{[2]}. For the same control system,different controller parameters may lead to totally different control results,if we choose the controller parameters according to the stability theorem, carelessly. Especially when working conditions change,the current controller parameters may cause the output response oscillation.
How to get appropriate and robust controller parameters is very important in real application. Although the MFAC is a pure data driven control method which is independent of the process model,it would be helpful to the parameter tuning if the first principle model or identified model is available. Even though it is not an accurate mathematical model,it still can be used to tune the parameters of MFAC to improve the control performance of the closed loop system. The parameter tuning of an improved MFAC without considering noise was discussed preliminarily in [6] where the optimal controller parameters were obtained for several different objective functions by using the approximate identified model. It was shown that the association performance index of squareerror and squareinput with raise time constraint (which is called Jeu_tr index) makes the system possess better dynamic response,and less iteration times,compared with other indices,such as integral of the squared error (ISE) and integral of the timeweighted absolute error (IAE).
In addition,the external stochastic noise has a great influence on the control system in the industrial filed. Sometime,it is not the Gaussian noise. Generally,the minimized variance of the tracking error can indicate optimization of controller tuning for systems which have Gaussian noise or at least have symmetric probability density functions^{[7]}. It is still not easy to find the suitable controller parameters during the finite time to guarantee the closeloop system is stable in the high noise environment,due to the limit of modern optimization method. Furthermore,for nonGaussian systems,the variance of tracking error cannot be used to represent the performance of the closedloop system alone. The reason is simple: if the interference is nonsymmetric,its spread area cannot be purely described by the variance. As a result,a more general measure of uncertainty,namely entropy,should be used to characterize the uncertainty of the tracking error^{[8]}. This advantage is especially significant in dealing with problems involving nonGaussian or large Gaussian statistics,where mean square methods may be inadequate. The application of entropy originates from the information theory^{[9]}. Various types of entropy measures have been introduced to control the stochastic nonlinear system^{[10]}.
In this paper,the parameters tuning problems of dynamic linearization based MFAC are considered. First,the dynamic linearization based MFAC algorithm with control input linearization length constant $L$ = 2 is rederived using the concept of pseudo gradient proposed in [2]. Then for the unavoidable problem of parameter tuning of stochastic nonlinear system,a minimum entropy performance index is proposed. Combined with the gradient descent technique,the parameter$'$s optimization method is derived. There are a few parameters in the MFAC algorithm,so it is difficult and unnecessary to tune all the parameters at one time. In this paper, the tuning of three key parameters is considered,and a simple offline practical tuning algorithm is given,in which a simplified and approximated model is used. At last,simulation examples for linear and nonlinear systems are given. In the case of Gaussian stochastic noise and nonGaussian stochastic noise,the dynamic performance of closeloop system based on entropy index tuning parameters has been compared with that of the traditional index, such as ISE and ITAE.
Ⅱ. MODEL FREE ADAPTIVE CONTROLA general discrete time nonlinear system is given as follows:
\begin{align} y(k + 1) = f(y(k),\cdots ,y(k  m),u(k),\cdots ,u(k  n)), \label{eq1} \end{align}  (1) 
In a traditional modelfree adaptive controller,a nonlinear system is transformed into a linear timevarying dynamic system according to the compact format linearization. Then the input and output relationship can be described as follows:
\begin{align} y(k + 1)  y(k) = \varphi (k)[u(k)  u(k  1)], \end{align}  (2) 
At first,assume that system (1) satisfies the generalized Lipschitz conditions^{[2]},that is
\begin{align} \left {y({k_1} + 1){\rm{  }}y({k_2} + 1)} \right \le b\left {{u_L}({k_1})  {u_L}({k_2})} \right, \end{align}  (3) 
When the nonlinear system (1) satisfies the above conditions, bounded $\phi_1(k)$ and $\phi_2(k)$ must exist such that system (1) can be denoted as the following partial form dynamic linearization (PFDL) data model:
\begin{align} \begin{split} y(k + 1)&  y(k) = {\varphi _1}(k)[u(k)  u(k  1)]+\\ &{\varphi _2}(k)[u(k  1)  u(k  2)], \end{split} \label{eq4} \end{align}  (4) 
\begin{align} \Delta y(k + 1) = {\varphi _1}(k)\Delta u(k) + {\varphi _2}(k)\Delta u(k  1). \end{align}  (5) 
It is well known that the system output at the current time is always affected by the system inputs at the current time and previous time. Therefore,(4) or (5) represents more input information about the controlled plant than (2).
An expanded integrated square error index (6) is used to solve the reference trajectory tracking problem.
\begin{align} J(u(k)) = {[{y^*}(k + 1)  y(k + 1)]^2} + \lambda {[u(k)  u(k  1)]^2}, \end{align}  (6) 
According to gradient descent technique,the control law is
\begin{align} \begin{split} u(k) &= u(k  1)+\\ & \frac{{{\rho _1}{\varphi _1}(k)[{y^*}(k + 1)  y(k)]}}{{\lambda + \varphi _{_1}^2(k)}}  \frac{{{\rho _2}{\varphi _1}(k){\varphi _2}(k)\Delta u(k  1)}}{{\lambda + \varphi _{_1}^2(k)}}, \end{split} \label{eq7} \end{align}  (7) 
\begin{align} \begin{split} &J({\varphi _1}(k),{\varphi _2}(k))= [y(k)  y(k  1)  {\varphi _1}(k)\Delta u(k  1)+\\ &\quad {\mu _1}{[{\varphi _1}(k)  {\hat{\varphi}_1}(k  1)]^2}+ {\mu _2}{[{\varphi _2}(k)  {\hat{\varphi} _2 }(k  1)]^2}, \end{split} \label{eq8} \end{align}  (8) 
\begin{align} &{{\hat \varphi }_1}(k)= {{\hat \varphi }_1}(k  1)+ \frac{{{\eta _1}\Delta u(k  1)}}{{{\mu _1} + \Delta {u^2}(k  1)}}\times\notag\\ &\quad [\Delta y(k)  {{\hat \varphi }_1}(k  1)\Delta u(k  1)  {{\hat \varphi }_2}(k  1)\Delta u(k  2)],\\ \end{align}  (9) 
\begin{align} &{{\hat \varphi }_2}(k) = {{\hat \varphi }_2}(k  1)+ \frac{{{\eta _2}\Delta u(k  2)}}{{{\mu _2} + \Delta {u^2}(k  2)}}\times\notag\\ &\quad [\Delta y(k)  {{\hat \varphi }_2}(k  1)\Delta u(k  2)  {{\hat \varphi }_1}(k  1)\Delta u(k  1)], \label{eq10} \end{align}  (10) 
As shown in the control law (7),(9) and (10),this kind of controller has nothing to do with the model structure or the order of the system. It is designed only by the input and output data.
Ⅲ. PARAMETERS TUNING BASED ON MINIMUM ENTROPY OPTIMIZATIONThe research about parameter tuning of the MFAC is of great significance to the applications of MFAC in practice engineering. There are a few parameters that need to be tuned for the MFAC (7), (9),and (10),and the suitable parameters are not easy to be obtained.
Firstly,the adjustable parameters of a second order partial format adaptive controller are $\rho_1$,$\rho_2$,$\lambda$ , $\eta_1$,$\eta_2$,$\mu_1$ and $\mu_2$. After a lot of engineering practices and simulation experiments,it is found that the key parameters are $\lambda$,$\rho_1$,and $\rho_2$. Parameter $\rho_2$ just appears in the second order partial format. Generally,the range of $\rho_2$ is (0,2) and $\lambda$ can be determined according to [2]. When $\rho_2/ \rho_1$ is constant and an appropriate $\lambda$ given,$\rho_1$ will be the exclusive parameter that should be tuned.
A. General Noise CaseSince the stochastic noise exists in the unknown plant (1),a minimum entropy index is proposed. Entropy is a unified probabilistic measure of uncertainty quantification. The entropy of a continuous type random error variable $x$ is defined as^{[14]}
\begin{align} H(x) =  \int\nolimits_{  \infty }^\infty {\gamma (x)} \ln \gamma (x){\rm d}x,\label{eq11} \end{align}  (11) 
The MFAC is an effective approach for actual systems without accurate model information,and the control effect is determined by control parameters. Generally,we can select appropriate parameters according to engineering experience or a trail and error method. Reference [2] gives a general parameter rule and stable parameter domain. But the system model information,if available,would be helpful for obtaining quantitative parameter descriptions for an endless variety of processes. This is the same as the PID parameter tuning,such as ZN,ISE and ITAE indices,even though PID is the most classical datadriven controller. That means controller design and parameter tuning are two different problems,and the model information is also necessary for the quantitatively parameter tuning problem of modelfree controller. Here we suppose that system model is known or identified from actual process data. The model for parameter tuning may be not as accurate as what we need for controller design.
Assume that the approximate or identified system model can be represented by the following nonlinear ARMAX equation:
\begin{align} \begin{split} y(k + 1) = f(y(k),\cdots ,y(k  m),\\ u(k),\cdots ,u(k  n),\omega (k)), \end{split} \label{eq12} \end{align}  (12) 
The output tracking error is defined as
\begin{align} \begin{split} &e(k) = y(k)  {y^*}(k)=\\ &\quad f(y(k  1),\cdots ,y(k  m),\\ &\quad u(k),\cdots u(k  n),\omega (k))  {y^*}(k):=\\ &\quad g(y(k  1),\cdots ,y(k  m),\\ &\quad u(k),\cdots ,u(k  n),\omega (k),{y^*}(k)), \end{split} \label{eq13} \end{align}  (13) 
\begin{align} &\eta (k) = \notag\\ &\quad (y(k  1),\cdots ,y(k  m),u(k  1) \cdots u(k  n),{y^*}(k)), \end{align}  (14) 
\begin{align} e(k) = g(\eta (k),u(k),\omega (k)). \end{align}  (15) 
Obviously,(15) is a general form of the tracking error. Since $g(\cdot)$ is a bounded function and $\omega(k)$ is defined on $a,b$,$e(k)$ will also be bounded if $u(k)$ is bounded. The tracking error $e(k)$ is not necessarily Gaussian distribution because $\omega(k)$ is a general stochastic noise. Therefore,it is not appropriate to simply use variance to characterize its randomness. Here entropy is employed to represent the uncertainty of the tracking error. In this case,the probability density functions of $e(k)$ should be first established.
Using the probability theory^{[8]},the probability density functions of $e(k)$ can be formulated as
\begin{align} \begin{split} {\gamma _{e(k)}}(\eta (k),u(k),x) = {\gamma _\omega }({g^{  1}}(\eta (k),u(k),x))\times\\ \quad \frac{{{\rm d}({g^{  1}}(\eta (k),u(k),x))}}{{{\rm d}x}},x \in [{a_e},{b_e}], \end{split} \label{eq16} \end{align}  (16) 
It can be seen from (16) that the shape of the probability density functions of the tracking error at time $k$ is determined by the controller output $u(k)$. Generally,the controller design should be performed such that the probability density function of the tracking error is as narrow as possible. This is simply because a narrow distribution function generally indicates that the uncertainty of the related random variable is small,which also corresponds to a small entropy value. According to the analysis of the above paragraph,parameter $\rho_1$ has a great effect on the controller output $u(k)$. For this purpose,the performance function (11) is rewritten as
\begin{align} J({\rho_1},&{\rho _2},\lambda ) =  \int\nolimits_0^N {\int\nolimits_{a_e}^{b_e} {{\gamma _{e(k)}}(\eta (k),u(k),x)}}\times\notag\\ &\ln {\gamma _{e(k)}}(\eta (k),u(k),x){\rm d}x{\rm d}t, \end{align}  (17) 
The performance function is a general nonlinear function with respect to $\rho_{1},\rho_2,\lambda$. Sometimes it is not necessarily convex. Therefore,the tuning of parameter $\rho_1$ is a nonlinear optimization problem that can be solved using the following gradient algorithm:
\begin{align} {\rho _1}(i + 1) = {\rho _1}(i)  b \frac{{\partial J}}{{\partial {\rho _1}(i)}}, \end{align}  (18) 
\begin{align} \begin{split} \frac{{\partial J}}{{\partial {\rho _1}}} =  \int\nolimits_0^N {\int\nolimits_{a_e}^{b_e} {\frac{{\partial {\gamma _{e(k)}}(\eta (k),u(k),x)}}{{\partial {\rho _1}}}} }\times\\ \quad (\ln {\gamma _{e(k)}}(\eta (k),u(k),x) + 1){\rm d}x{\rm d}t \end{split} \label{eq19} \end{align}  (19) 
Here we just give the tuning of parameters $\rho_1$. The tuning of parameter $\lambda$ and $\rho_2$ is similar. Actually,the performance function (17) is a function of parameters $\lambda$, $\rho_1$ and $\rho_2$. Using the gradient algorithm,we can get the optimal values of $\lambda$ and $\rho_2$. \begin{align*} {\rho _2}(i + 1) = {\rho _2}(i)  c \frac{{\partial J}}{{\partial {\rho _2}(i)}}, \end{align*} \begin{align*} \lambda (i + 1) = \lambda (i)  d \frac{{\partial J}}{{\partial \lambda (i)}}, \end{align*} with \begin{align*} \begin{split} \frac{{\partial J}}{{\partial {\rho _2}}} =  \int\nolimits_0^N {\int\nolimits_{a_e}^{b_e} {\frac{{\partial {\gamma _{e(k)}}(\eta (k),u(k),x)}}{{\partial {\rho _2}}}} }\times\\ \quad (\ln {\gamma _{e(k)}}(\eta (k),u(k),x) + 1){\rm d}x{\rm d}t, \end{split} \end{align*} \begin{align*} \begin{split} \frac{{\partial J}}{{\partial \lambda }} =  \int\nolimits_0^N {\int\nolimits_{a_e}^{b_e} {\frac{{\partial {\gamma _{e(k)}}(\eta (k),u(k),x)}}{{\partial \lambda }}} }\times\\ \quad (\ln {\gamma _{e(k)}}(\eta (k),u(k),x) + 1){\rm d}x{\rm d}t. \end{split} \end{align*}
B. Linear Gaussian CaseThe Gaussian stochastic noise is a special case of general stochastic noise. It has symmetrical probability density,and the parameter is optimal when the error is the smallest value. The expression of the minimum entropy optimization algorithm is rederived by the probability density of Gaussian stochastic noise.
Consider the following CARMAX model widely used in stochastic and adaptive control:
\begin{align} y(k) = \sum\limits_{j = 1}^m {{a_j}y(k  j)} + \sum\limits_{l = 0}^n {{b_l}u(k  l) + \omega (k)}, \end{align}  (20) 
\begin{align} {\gamma _\omega }(x) = \frac{1}{{\sqrt {2\pi } \sigma }}{{\rm e}^{  \frac{{(x  \mu )}^2}{2{\sigma ^2}}}},x \in (  \infty ,+ \infty ). \end{align}  (21) 
This system is a special case of the general system (12) and its tracking error takes the following simple form:
\begin{align} \begin{split} e(k) &= y(k)  {y^*}(k)= \sum\limits_{j = 1}^m {{a_j}y(k  j)} +\\ & \sum\limits_{l = 1}^n {{b_l}u(k  l) + {b_0}u(k) + \omega (k)  {y^*}(k)}. \end{split} \label{eq22} \end{align}  (22) 
Accordingly,the known vector $\eta(k)$ in (15) is grouped by
\begin{align} \eta (k) = \sum\limits_{j = 1}^m {{a_j}y(k  j)} + \sum\limits_{l = 1}^n {{b_l}u(k  l){\rm{  }}{y^{\rm{*}}}(k)}. \end{align}  (23) 
\begin{align} \begin{split} e(k) &= \eta (k) + {b_0}u(k) + \omega (k):=\\ & g(\eta (k),u(k),\omega (k)). \end{split} \label{eq24} \end{align}  (24) 
\begin{align} \begin{split} \omega(k)&= e(k){\rm{  }}\eta (k)  {b_0}u(k)=\\ & {g^{  1}}(\eta (k),u(k),x), \end{split} \label{eq25} \end{align}  (25) 
\begin{align} {\gamma _{e(k)}}(\eta (k),u(k),x) = \frac{1}{{\sqrt {2\pi } \sigma }}{{\rm e}^{  \frac{{(x  \eta (k)  {b_0}u(k)  \mu )}^2}{2{\sigma ^2}}}}. \end{align}  (26) 
\begin{align} &\frac{{\partial {\gamma _{e(k)}}(\eta (k),u(k),x)}}{{\partial {\rho _1}(i)}} = \frac{{\partial {\gamma _{e(k)}}(\eta (k),u(k),x)}}{{\partial u(k)}}\times \frac{{\partial u(k)}}{{\partial {\rho _1}(i)}},\notag\\ &\frac{{\partial {\gamma _{e(k)}}}}{{\partial u}} = \frac{{{b_0}}}{{2{\sigma ^3}\sqrt {2\pi } }} {{\rm e}^{  \frac{{(x  \eta (k)  {b_0}u(k)  \mu )}^2}{2{\sigma ^2}}}}, \label{eq27} \end{align}  (27) 
\begin{align} \begin{split} &\ln {\gamma _{e(k)}}(\eta (k),u(k),x)=\\ &\quad \ln \frac{1}{{\sqrt {2\pi } \sigma }}  \frac{1}{{2{\sigma ^2}}}{(x  \eta (k)  {b_0}u(k)  \mu )^2}. \end{split} \label{eq28} \end{align}  (28) 
\begin{align} \begin{split} \frac{{\partial J}}{{\partial {\rho _1}}} &=  \int\nolimits_0^N {\int\nolimits_{{a_e}}^{{b_e}} {\frac{{{b_0}}}{{{\sigma ^3}\sqrt {2\pi } }} {{\rm e}^{  \frac{{(x  \eta (k)  {b_0}u(k)  \mu )}^2}{2{\sigma ^2}}}}}} \times\\ &\frac{{{\varphi _1}(k)[{y^*}(k + 1)  y(k)]}}{{\lambda + \varphi _{_1}^2(k)}} \times\\ &\left(\ln \frac{1}{{\sqrt {2\pi } \sigma }}  \frac{1}{{2{\sigma ^2}}}{(x  \eta (k)  {b_0}u(k)  \mu )^2} + 1\right){\rm d}x{\rm d}t \end{split} \label{eq29} \end{align}  (29) 
The parameter tuning process of dynamic linearization based MFAC is shown in the following:
1) Identify a simple model for system (12) or (20) according to offline data,which can be not very accurate.
2) Select an appropriate performance index $J$ in (17) according to the control requirement.
3) Initialize other parameters,such as $b$,$\rho_2$,$\lambda$, $\eta_1$,$\eta_2$,$\mu_1$ and $\mu_2$.
4) Give an initial parameter $\rho_1$ in the range of (0,2).
5) Compute the gradient function of $J$ with respect to parameter $\rho_1(i)$,that is,$\frac{\partial J}{\partial {\rho _1}}(i)$.
6) Update the tuning parameter $\rho_{1}$ according to \begin{align*} {\rho _1}(i + 1) = {\rho _1}(i) + \Delta {\rho _i}, \end{align*} where $\Delta {\rho _1} =  b \cdot \frac{\partial J}{\partial {\rho _1}(i)}$.
7) If the change of index $J$ meets $\Delta J < \varepsilon $,the parameter tuning is over; else jump to step 5) and keep tuning.
The procedure of tuning is shown in Fig. 1.
Download:


Fig. 1. Flow chart of a practical tuning algorithm. 
In this section,two different systems are used to test the availability and effectiveness of the proposed entropy tuning method. They are a thirdorder linear system with large delay and a timevarying nonlinear system. In order to analyze the effectiveness of the minimum entropy parameter tuning method under random noise,the other two classic performance indexes are given for comparison. The optimum PID controller algorithm was introduced by [16]. And the more general index is
\begin{align} J = {\int_0^\infty {\left[{{t^n}e(t)} \right]} ^2}{\rm d}t, \label{eq30} \end{align}  (30) 
In the parameter optimization,ITAE is often used as the performance index^{[18]},and it is described as
\begin{align} J = \int_0^\infty {te(t)} {\rm d}t. \end{align}  (31) 
For the Gaussian and nonGaussian stochastic noise,the parameter of MFAC (7),(9),(10) is optimized by three performance indices, i.e.,ISE,ITAE,and minimum entropy index.
A. Example of Linear System with Large DelayGiven a thirdorder large inertial delay system^{[19]}:
\begin{align} G(s) = \frac{{K{{\rm e}^{  \tau s}}}}{{{{\left( {1 + Ts} \right)}^3}}}, \end{align}  (32) 
In engineering practice,it is common that any higher order and complex objectives can be approximated by a secondorder inertial model plus time delay. Suppose that the first principle model and the identified model are available as shown by (32) and (33), respectively. We can see from Fig. 2 that it is not an accurate mathematical model. In the following,we will show that it still can be used to tune the parameters of MFAC to improve the control performance of the closed loop system. The result of parameter tuning is accepted even the model identification error is very large.
Download:


Fig. 2. Output responses of system and simple model. 
Suppose that the approximate model is
\begin{align} G(s) = \frac{{K{{\rm e}^{  \tau s}}}}{{{{\left( {1 + Ts} \right)}^2}}}, \end{align}  (33) 
Parameter Optimization Effect on the System Stability
i) The case of Gaussian stochastic noise
During the simulation experiment,50 groups of 2.5 % Gaussian stochastic noise are arbitrarily selected. The probability distribution of the Gaussian stochastic noise refers to (21). In order to test and verify the effectiveness of different parameter optimization criteria,every performance index is implemented under the same disturbance.
After 50 groups of Gaussian stochastic noise disturbance experiments,50 groups of different parameter $p_1$ are got. First,it is considered whether the step output response of closeloop system (32) and (7) with the tuned parameter is stable or not. The results are listed in Table I.
As shown in Table I,the number of stable responses under the minimum entropy index is the highest. From the tuning result,we know the optimal parameter is near 0.003,so the initial values $\rho_{1,0}$ = 0.01 and 0.001 represent the two different optimizing directions. It is shown that 50 groups parameter obtained from entropy criterion are all appropriate and can get stable responses even if the optimization is from a different direction. Whereas the results obtained from the traditional indexes,such as ISE and ITAE,are not encouraging. It seems that these two criterions are not very adaptive to the big noise situation. The reason is obvious that the existing stochastic noise is considered in designing of minimum entropy optimization but not in ISE and ITAE methods. Controller parameter is optimized according to the effect degree of stochastic noise,therefore,the minimum entropy algorithm can guarantee the controlled system stable under any Gaussian stochastic noise.
ii) The case of nonGaussian stochastic noise
Suppose that the probability distribution of the nonGaussian stochastic noise is
\begin{align} \gamma (x) = \left\{ {\begin{array}{*{20}{lll}} \frac{3}{{4\sqrt 5 }}(1  0.2{x^2}),&\left x \right \le \sqrt 5 ,\\ 0,&{\rm otherwise}. \end{array}} \right. \label{eq34} \end{align}  (34) 
\begin{align} \eta = \frac{\mbox{number of stable response}}{\mbox{total test number}}. \end{align}  (35) 
As shown in Table II,the result of stability simulation is similar to the Gaussian noise case. The results obtained from the traditional indexes,such as ISE and ITAE,are worse than those of the entropy indexes. It is noted that in the nonGaussian noise experiment,the stable degrees of ISE and ITAE drop down significantly,while the stable degree of the entropy is the same as the Gaussian noise experiment. The minimum entropy index has better performance in the case of Gaussian and nonGaussian noise.
It can be shown from Table III that the stable ratio of nonGaussian noise is much bigger than that of Gaussian noise for traditional indices ISE and ITAE. Furthermore,the traditional performance index cannot completely satisfy the stability requirement under the nonGaussian stochastic noise circumstance. However,the minimum entropy tuning algorithm can guarantee the system stable in the both cases of Gaussian and nonGaussian noise.
In order to ensure the controlled system is stable under different performance indices,the stochastic noise is decreased as 1 %. The parameter $\rho_1$ tuning process from the two directions ($\rho_{1,0}$ = 0.01 or $\rho_{1,0}$ = 0.001) are shown in Fig. 3. Fig. 4 shows the performance index changing with iterations. Relatively,the entropy method has a more rapid convergence rate and more stable optimal results than those of the ITAE and ISE indices.
Download:


Fig. 3. Parameter $\rho_1$ tuning process. 
Download:


Fig. 4. ISE, ITAE, and entropy index changing when initial $\rho_{1,0}$ = 0.001. 
1) Control Results Under Different Performance Index Parameter Optimization
In the case of the nonGaussian stochastic noise ($\pm1$ %),the initial parameters are $\rho_{1,0}$ = 0.01,$\rho_{1,0}$ = 0.001, the parameter $\rho_{1}$ of the dynamic linearization MFAC is tuned using ISE,ITAE,and minimum entropy optimization indices, respectively. The system output respond and controller output under different performance indexes are shown in Figs. 5 and 6. Even under the conditions that system noise is small and the ISE or ITAE method can obtain desired parameters,the control response under the entropy index is very good,which has more rapid response than that of ISE index. The dynamic response results of three performance indices are listed in Table IV.
Download:


Fig. 5. System and controller output with initial $\rho_{1,0}$ = 0.01. 
Download:


Fig. 6. System and controller outputs with initial $\rho_{1,0}$ = 0.001. 
As shown in Table IV,the dynamic response of ISE is too slow in the case of the nonGaussian noise. Besides,the overshoot of the ISE method is still very large. The results of the ITAE and entropy are approximate. But in the view of the rising time, entropy has a better response speed. Therefore,minimum entropy makes the system have a better dynamic performance,which has the less rising time,the less setting time,the smaller overshoot value and so on.
The following experiment is designed to change the system parameter of the controlled objective in order to test the robustness of controller parameter tuned by ISE,ITAE and Entropy methods,respectively. Here the system parameters are changed as follows: \begin{align*} \left\{ {\begin{array}{*{20}{l}} {K = 5,T = 3,\tau = 60,0 < t \le 1 500}\\ {K = 4.5,T = 0.8,\tau = 50,1 500 < t \le 3 000}\\ {K = 5,T = 3,\tau = 60,3 000 < t \le 4 500} \end{array}} \right. \end{align*}
At the time of $1 500 < t < 3 000$, system parameters,such as openloop gain $K$,time constant $T$ and delay time all change. The system step response and controller output are given in Fig. 7,in which controller parameters are selected as the same as Table IV. It is shown that system responses are not affected by the changing of system. That means the entropy method has good robustness.
Download:


Fig. 7. Control results with system parameters changes: $\rho_{1,0}$ = 0.01. 
The similar experiment is designed to verify the robustness and dynamic stability of a nonlinear system. Three parameter optimization indices,ISE,ITAE and entropy are compared.
Given a nonlinear system,which consists of two nonlinear subsystems^{[20]} \begin{align*} &y(k + 1) =\\[2mm] &\quad \left\{ {\begin{array}{*{20}{lllll}} \frac{{y(k)}}{{1 + y{{(k)}^2}}} + u{{(k)}^3},&k \le 500,\\ \frac{{y(k)y(k  1)y(k  2)u(k  1)(y(k  2)  1) + a(k)u(k)}}{{1 + y{{(k  1)}^2} + y{{(k  2)}^2}}},&k > 500, \end{array}} \right. \end{align*} where $a(k)=1+{\rm round}(k/500)$,the second subsystem has a step timevarying parameter $a(k)$. Therefore,the structure,order and parameter of this system are timevarying. The openloop step response of the nonlinear system (35) is shown in Fig. 8 (a). Fig. 8 (b) is the change process of the timevarying parameter $a(k)$. It is shown that the openloop response has great oscillation at 500 s and 750 s,and the steady output value changes with time. That is because the structure of controlled objects and timevarying parameter $a(k)$ change with time. Here the approximate model (33) is still used for controller parameter tuning.
Download:


Fig. 8. Nonlinear system output response. 
In the case of the nonGaussian stochastic noise ($\pm$ 4 %), the initial parameter is $\rho_{1,0}$ = 0.8. Parameter $ \rho_{1}$ is tuned using ISE,ITAE and minimum entropy optimization methods,respectively. Here given other parameters are $\lambda$ = 2,$\rho_2$ = 0.35,$\eta_1$ = 1,$\eta_2$ = 1, $\mu_1$ = 1 and $\mu_2$ = 2.
After tuning by ISE,ITEA,and entropy indices,we selected optimal parameters as $\rho_1$ = 0.16888 (ISE),$\rho_1$ = 0.19897 (ITAE),$\rho_1$ = 0.27689 (entropy). The closeloop system responses under different parameters are given in Fig. 9. The controlled object is the first subsystem at the time of $0\leq t\leq 500$,in which timevarying parameter $a(k$) has no influence on it. Then its output response is rapidly tracking the setpoint $y^{\ast}$. However,at the time of $500\leq t\leq 1 000$,the controlled object is the second subsystem. Therefore, the system output is affected by timevarying parameter changing at 750 s. Even the structure,order and parameter of the controlled object are timevarying,the closeloop system response obtained by the entropy index shows stronger robustness,rapid convergence rate and accurate tracing result.
Download:


Fig. 9. Output responses for nonlinear system: $ \rho_{1,0}$ = 0.8. 
The MFAC method is considered,in which the system input and output relationships are described as the partial form linearization data model. Then the minimum entropy idea is used to solve the parameter optimization problem of the MFAC in the case of Gaussian noise and nonGaussian noise. According to the probability distribution of stochastic noise,the entropy recursive optimization algorithm is derived in the case of general random noise. By choosing a lot of Gaussian stochastic noise and nonGaussian stochastic noise,the universality of minimum entropy optimization algorithm is verified. In comparison with the other traditional performance indices,such as ISE and ITAE,it is found that: the existing stochastic noise has been considered in the designing of minimum entropy index,so the entropy optimization algorithm can guarantee the controlled system stable under Gaussian or nonGaussian stochastic noise. Meanwhile the optimal parameter obtained from the minimum entropy index can assure that the system has better dynamic performance,such as less rising time,less stability time and smaller overshoot.
[1]  Hou Z S, Huang W H. The modelfree learning adaptive control of aclass of SISO nonlinear systems. In: Proceedings of the 1997 AmericanControl Conference. New Albuquerque, NM: IEEE, 1997.343344 
[2]  Hou Z S, Jin S T. Model Free Adaptive Control: Theory and Application.Boca Raton: CRC Press, 2013. 
[3]  Hou Z S, Jin S T. Data driven modelfree adaptive control for a classof MIMO nonlinear discretetime systems. IEEE Transactions on NeuralNetworks, 2011, 22(12): 21732188 
[4]  Xu D Z, Jiang B, Shi P. A novel model free adaptive control designfor multivariable industrial processes. IEEE Transactions on IndustrialElectronics, to be published 
[5]  Hou Z S, Wang Z. From modelbased control to datadriven control:survey, classification and perspective. Information Sciences, 2013, 235:335 
[6]  Wang J, Ji C, Cao L L, Jin Q B. Model free adaptive control and parametertuning based on second order universal model. Journal of CentralSouth University (Science and Technology), 2012, 43(5): 17951802 
[7]  Astrom K J. Introduction to Stochastic Control Theory. New York:Academic, 1970. 
[8]  Yue H, Wang H. Minimum entropy control of closedloop trackingerrors for dynamic stochastic systems. IEEE Transactions on AutomaticControl, 2003, 48(1): 118122 
[9]  Deniz E, Jose C. An errorentropy minimization algorithm for supervisedtraining of nonlinear adaptive systems. IEEE Transactions on SignalProcessing, 2002, 50(7): 17801786 
[10]  Petersen I R, James M R, Dupuis P. Minimax optimal control ofstochastic uncertain systems with relative entropy constraint. IEEETransactions on Automatic Control, 2000, 45(3): 398412 
[11]  Ma Y, Chen X, Xie X H. Research on MFA control algorithm withtracking differentiator. Chinese Journal of Scientific Instrument, 2009,30: 204208 
[12]  Chi R H, Hou Z S. A model free periodic adaptive control for freewaytraffic density via ramp metering. Acta Automatica Sinica, 2010, 36(7):10291033 
[13]  Coelho L S, Pessôa M W, Sumar R R. Model free adaptive controldesign using evolutionaryneural compensator. Expert Systems withApplications, 2010, 37(1): 499508 
[14]  Yue H, Zhou J L, Wang H. Minimum entropy of Bspline PDF systemswith mean constraint. Automatica, 2006, 42(6): 989994 
[15]  Papoulis A, Pillai S U. Probability, Random Variables, and StochasticProcesses. New York: McGrawHill, 1991. 
[16]  Zhuang M, Atherton D P. Tuning PID controllers with integral performancecriteria. In: Proceedings of the 1991 International Conference onControl.Edinburgh: IEEE, 1991.481486 
[17]  Zhuang M, Atherton D P. Automatic tuning of optimum PID controllers.IEE Proceedings D, Control Theory and Applications, 1993, 140(3):216224 
[18]  Xu F, Li D H, Xue Y L. Comparing and optimum seeking of PIDtuning methods base on ITAE index. Proceedings of the Csee, 2003,23(8): 206210 
[19]  Jin S T, Hou Z S. An improved modelfree adaptive control for a classof nonlinear largelag systems. Control Theory & Applications, 2008,25(4): 623626 
[20]  Narendra K S, Parthasarathy K. Identification and control of dynamicalsystems using neural networks. IEEE Transaction on Neural Networks,1990, 1(1): 427 