The approximate nonlinear receding-horizon control law is used to treat the trajectory tracking control problem of rigid link robot manipulators. The derived nonlinear predictive law uses a quadratic performance index of the predicted tracking error and the predicted control effort. A key feature of this control law is that, for their implementation, there is no need to perform an online optimization, and asymptotic tracking of smooth reference trajectories is guaranteed. It is shown that this controller achieves the positions tracking objectives via link position measurements. The stability convergence of the output tracking error to the origin is proved. To enhance the robustness of the closed loop system with respect to payload uncertainties and viscous friction, an integral action is introduced in the loop. A nonlinear observer is used to estimate velocity. Simulation results for a two-link rigid robot are performed to validate the performance of the proposed controller.
During recent years much emphasis has been placed on flexible manufacturing processes where the most important factors are quality, costs and time. Both fast motion in unconstrained space and mechanical interaction with the environment are required in most manufacturing systems. Industrial robots are often used to meet this demand and to perform various tasks such as material assembling, painting or welding. To accomplish these tasks efficiently and accurately, several control approaches have been proposed in the literature. Among these, a simple PD-control scheme that achieves satisfactory performance (Spong M. W. & Vidyasagar M. 1989) in the absence of gravity. However, robot manipulator is highly nonlinear system with coupling between joints and the gravity effects. The computed torque control or feedback linearization control has been also used to achieve best tracking performance. The implantation of the computed torque controller requires exact knowledge of the robot dynamics. Unfortunately, model uncertainties are frequently encountered in robotics due to unknown or changing payload and friction. These model uncertainties may decrease significantly the performance of this method in terms of tracking accuracy. Therefore, to achieve acceptable performance, even when all kinds of uncertainties are encountered, numerous robust control algorithms have been used like the variable structure approach (Slotine J. J. E. & Sastry S. S. 1983), robust adaptive approach (Ortega R. & Spong M. W. 1989), (Lee K. W. & Khalil H. K. 1997), (Canudas C. W. & Fixot N. 1992), (Spong M. W. 1992) and nonlinear H∞ approach (Chen B. S. et al 1994). A first survey of early results in robust control has been compiled in (Abdullah C. et al; 1991) and the second survey of recent results has been given in (Sage H. G. et al, 1999). Finally the robotic applications require effective control laws that achieve accurate tracking of fast motion despite the variations of inertia and gravitational load of the manipulator during operation.
Model predictive control of linear systems has received considerable attention in the last decade due to its robustness with respect to model uncertainties. However, many systems are inherently nonlinear. Since linear models are often inadequate to describe accurately the process dynamics, then nonlinear models should be used. Much effort has been made to extend linear predictive control to nonlinear systems (Michalska H & Mayne D. Q. 1993). The disadvantages of the proposed approach is the heavy online computation burden that causes two important problems in implementation of the nonlinear predictive control. One is the computation delay that cannot be ignored and the other is the global solution that cannot be guaranteed in each optimization problem. The application of these nonlinear control laws to nonlinear systems characterized by fast dynamics (such robotics) sound like unusual proposal. To overcome the computation burden, several nonlinear predictive laws have been developed in (Ping L. 1995), (Singh S. M. 1995), (Souroukh M. & Kravaris C. 1996), (Chen W. H. et al, 2003), where the one step ahead predictive output error is obtained by expanding the output signal and the is used to derive the offline control laws. In this paper, the nonlinear receding-horizon controller proposed in (Ping L. 1998) is applied to robot manipulator to achieve position angular tracking objectives. To derive the control law, the predictive tracking error and the predicted control effort are minimized over a fixed time horizon. This approximate nonlinear controller is given in a closed form and thus no online optimization is required. Moreover, to increase the robustness of the control algorithm with regard to model uncertainties, we propose to introduce an integral action in the loop. The well-known Lyapunov based theory is used to show the asymptotic stability of the closed loop system in matched or mismatched case.
The major drawback of the proposed schemes is the requirement of measurement of motor speed. Speed measurements increase cost and impose constraints on the achievable bandwidth. Thus, to overcome this problem a nonlinear observer is used to estimate position and velocity angular of robot manipulator.
The outline of this paper proceeds as follows. In the next section, a dynamic model of robot manipulator is presented. In section 3, the approximate receding-horizon control scheme is developed to allow position angular tracking of a desired references trajectory. Stability analysis and robustness are treated in section 4. The high gain observer used to estimate unmeasured output elements (velocity angular) is presented in section 5. Simulation results are given in section 6. In the last, we conclude with some remarks.
2. Dynamic model of rigid link robot manipulators
The Euler–Lagrange equations are a tool from analytical mechanics that can be used to derive the equations of motion for a mechanical system. In this approach the joint q(t) are considered as generalized coordinates. The kinetic energy of a robot manipulator with n degrees of freedom can be calculated as:
where D(q) is the inertia matrix. Let U(q) : ℜn → ℜ be a continuously differentiable function, called the potential energy. The Lagrangian function is defined (Spong M. W. & Vidyasagar M. 1989) by:
The dynamics of the manipulator are described by Lagrange's equations:
reference signal in a rith order Taylor series, ri is the relative degree of the ith element of the output. Then, the continuous minimization of the predictive tracking errors
where u1, u2,……un represent generalized input torques. Inserting the kinetic energy and the potential energy for the Lagrange L(q, q̇) above leads to the matrix description:
where q(t) ε ℜn is the vector of the generalized coordinates representing the angular joint positions and controlled with the driving torques ur ε ℜn, D(q) ε ℜnxn, D(q)=D(q)T>0, is the link inertia matrix, C(q, q̇)q̇ ε ℜn is the vector of the coriolis and centripetal torques, G(q) ε ℜn is the vector of gravitational torques and fr represents friction torques acting on the joints. This is described in [9], when only the mechanical parts of actuators dynamics are included. The dynamic model of a rigid robot manipulator becomes:
with:
and f = fr + Nfm,
where:
N is the diagonal matrix of the gear ratios.
um is the vector of torque supplied by the actuators.
fm is the vector friction torque acting on the motors.
Jm is the diagonal matrix containing the effective motors' inertia.
It is assumed that the position q(t) is available for measurement.
Control Objective: The desired reference trajectory for the control object to follow is assumed to be available as bounded functions of time in terms of generalized position qref (t). That is, there exist three positive constants ri, i=0,1,2 such that the following inequalities hold:
State space representation: The dynamic equation of n link robot manipulator (1) can be written in the state space representation as:
where x(t) = [x1x2]T = [qq̇]T ε ℜ2n is the state vector. u(t) ε ℜn represents the control torque vector and y(t) ε ℜn is the output vector (position angular). f(x1, x2) = –M(q)−1 (C(q, q̇)q̇ + G(q))) ε ℜn and P(x1) = M(x1)−1 ε ℜnxn.
P1. The matrix M(x1) is symmetric definite positive, then there exist two positive constants: M and M̄ such that: M ≤ ||M(x1)|| ≤ M̄.
P2. ∃ μ > 0 such that ||C(q, x)||≤ μ ||x||, ∀ x ε ℜn.
P3 The vector function f(x1,x2) is Lipschitz with respect to x2. Thus there exists κ > 0 such that:
3. Receding-horizon control law
In the receding-horizon control strategy, the following control problem is solved at each t>0 and x(t):
subject to the equation (3) and x(t + h) = 0 for some h>0, where Q is positive definite and R positive semi-definite. Denote the optimal control to the above problem by u*(τ), τ ε [t, t + h]. The currently applied control is u(t) set equal to u*(t). This process is repeated for every next t for stabilization of the system at the origin. However, to solve a nonlinear dynamic optimization problem with equality constraints is highly computationally intensive, and in many cases it is impossible to be performed within a reasonable time limit. Furthermore, the global optimization solution cannot be guaranteed in each optimization procedure since, in general, it is a non-convex, constrained nonlinear optimization problem.
In order to find the current control that improves tracking error along a fixed interval, the output tracking error e(τ) = q(τ) – qref(τ) is used instead of the state vector x(τ) in the above receding control problem:
where Qεℜnxn is positive definite, Rεℜnxn positive semi-definite, T is the predicted tracking horizon.
To avoid the computational burden, we shall approximate the above receding- horizon control problem by Simpson's rule (Atkinson K. E. 1978):
with T=2h is the prediction horizon and
A simple and effective way of predicting the cost function L(.) is to expand the predicted tracking error in a first order Taylor series, in the following way: q(t + h) = q(t) + hq̇(t) and the reference trajectory is predicted as follow: qref(t + h) = qref(t) + hq̇ref.
The predicted tracking error is then given by:
Predict e(t+2h) by another first-order Taylor series expansion at e(t+h) to have:
where and .
Thus, the performance index (5) can be approximated as:
We can rewrite the performance index (7) in the conventional quadratic form by using the predicted tracking error given above, as:
where is positive definite matrix, m(e, ė): terms that are independent of U(t) where U(t)T = [u(t)Tu(t+h)Tu(t+2h)T],
The receding-horizon control problem that minimizes the cost function J̄ is: U(t) = –θ(x)−1G(x). The applied control signal to nonlinear system at time t is given by:
Note that with R=0, the above nonlinear predictive control law leads to the well known computed torque controller.
4- Stability analysis and robustness issues
In this section, we will investigate the stability and the robustness of the closed loop system with respect to model uncertainties.
1- Stability Analysis
Let Q = qIn and R = rIn (we give the same penalty to all joints), the tracking error of the nonlinear system (3) closed by the nonlinear feedback (9) is given by:
Where and .
This equation can be written in compact form as:
Where ; ; .
Lemma-1: The matrix A(h,x1) is Hurwitz.
Proof: Both matrix P̄ and its inverse are symmetric positives definite. Let x̄ ε ℜn and are the eigenvector and the correspondent eigenvalue of the inverse of the above matrix. Thus, we have the equality:
ote that λ is the solution of equation:
Therefore, λ is the eigenvalue of the matrix A(h, x1) and the correspondent eigenvector. Set λ1 and λ2 the solution of the equation (12), we have the relations:
Since the eigenvalue is positive, then λ1 and λ2 have a negative real part (end of the proof).
Since the matrix A(h,x1) is a Hurwitz matrix, then for any symmetric positive definite matrix QA(h, x1), there exists a symmetric positive definite matrix PA(h, x1) solution of the lyapunov equation:
From the property P3, the function f(x1, x2) is Lipschitz with regards to x2, we can always find a bounded continuous function σ(e2, t) and positive scalar μ satisfying the inequality:
Now, we can state the following theorem.
Theorem 1: The equilibrium point of the nonlinear system (3) in closed loop with the feedback control (9) is asymptotically stable if the following inequality hold:
Moreover, if r = 0, then the origin is asymptotically stable equilibrium point.
Proof:
Let V(e) = eTPA e be a Lyapunov function candidate, the time derivative of this function along the trajectories (11) is:
Thus, V̇ ≤ –κ||e||2 is negative definite if κ > 0, where . This ensures the asymptotic stability of the equilibrium point.
Note that a short steady state error will be observed in tracking position error when r ≠ 0. However, if r = 0 the time derivative of the Lyapunov function becomes: V̇ = –λmin(QA)||e||2 which is negative definite. Thus, we can conclude that the origin becomes the equilibrium point of the system (12) and is asymptotically stable, i.e:
2- Robustness
In order to incorporate modeling uncertainties into the model of the rigid robot (1), the matrices M(q), C(q, q̇) and the vector G(q) are split up into a nominal part (indicated by the subscript zero) and an uncertain part as:
The friction torque f is included in the uncertain part given the difficulty to model it correctly. Obviously, only the nominal part of the model can be used by the nonlinear predictive control, given by:
Where R is setting to zero. With the nonlinear control law (16), the closed loop system is:
To estimate the worst case bound of the function υ, we make the following assumptions for all q ε ℜn:
Given these assumptions with the inequalities (2), we can find a bounded continuous vector function ρ(e, ė, t) satisfying the inequality (Spong M. W. & Vidyasagar M.; 1989):
where γ is a positive scalar and e = |qq̇|T = |e1e2|T.
In the state space representation, the system (17) can be transformed to:
where .
Since both M−1(x1) and M0(x1) are symmetric definite positives, the matrix b(x1) = M−1 (x1) M0 (x1) has all its eigenvalues reels and positives (Samson C.; 1983). Thus, from the Lemma-1, we can conclude that the matrix B̄(h, x1) is Hurwitz, then for any symmetric positive definite matrix QB(h,x1) there exists a positive definite matrix PB(h,x1) solution of the given Lyapunov equation:
Theorem 2: Suppose that the inequality holds , then the equilibrium point of the nonlinear system with uncertainties (15) closed by the optimal control (16) is asymptotically stable.
Proof:
Let V = eTPBe a Lyapunov function candidate. The differentiation of V along the trajectories (18) leads to:
Which is definite negative if . Therefore, By LaSalle's invariance theorem, the solution e(t) of (18) tends to the invariance set:
We conclude that bounded uncertainties will introduce a steady state error on tracking position angular. Where x̄ is the equilibrium point of the system (3) in closed loop with the control law (16).
3- Integral action
It is known in the literature that the integral action increases the robustness of the closed loop system against the low frequency disturbances as long as the closed loop system is stable. In this part, we shall incorporate an integral action in the loop to eliminate the steady state error and enhance the robustness of the proposed control scheme with respect to model uncertainties and disturbances.
Thus, the cost function to minimize becomes:
Where ė0= e1 = x1 – qref = q – qref and L(τ) = e0T(τ) Q e0 (τ) + uT(τ) R u (τ).
Also in this case, we use the Simpson's rule to approximate the integral in the cost function (19) by:
Note that in this case, the Taylor approximation of the predicted vector e0(t + h) is given by:
Following same steps in paragraph 2, the optimal control vector U(t) that minimizes the new cost function is:
Where U(t) = |u(t) u(t + h) u(t + 2h)|T; and
The control signal to be applied to the nonlinear system at time t is:
Where .
Let R = 0 in equation (21), the control signal becomes:
The dynamic of the tracking error is given by the equations:
Or in compact form:
Where:
Lemma 2: Suppose that λmax(ε(x1)) < 2.6, with , then the matrix is Hurwitz.
Proof: Let x̄ ε ℜn and represent eigenvector and eigenvalue of the matrix b(x1) respectively. Set be a vector and a scalar. We have:
Then, and v are the eigenvalue and eigenvector of the matrix B˜(h, x1) respectively, where the eigenvalue verify the equality:
By using the Rooth-Hurwitz criterion, the solutions of the above equation lie in left half plane (stable domain) if: .
Since M–1M0 = (In + ε(x1))–1, the condition for stability becomes: or λmax (ε(x1)) < 2.6 which can be easily verified if the uncertainty ΔM is small with regards to the nominal value of the matrix M0(x1).
We conclude that the matrix B˜(h, x1) is Hurwitz, then for any symmetric positive definite matrix Q˜(h, x1) there exists a positive definite matrix P˜(h, x1) solution of the given Lyapunov equation:
Theorem 3: Suppose the inequality hold, then the equilibrium point of the nonlinear system with uncertainties (23) is asymptotically stable.
The proof may be obtained in the same way as the proof of the theorem 2 and is therefore omitted. Therefore, the equilibrium point of (22) or (23) is asymptotically stable. The tracking error tends towards the set:
e conclude that the position and velocity tracking error converge to zero, therefore the integral action eliminates the position steady state error, i.e:
The price to be paid by introducing an integral action in the loop is that the control signal will not vanish and this will increase the required energy to maintain the tracking performance as in the matched case.
5. Nonlinear observer
A drawback of the previous nonlinear predictive controller is that it requires at least the measurement of velocity on the link side. However, as pointed out in (Nicosia S. & Tomei P. 1990) and (Canudas W. C. et al 1992), in the practical robotic systems all the generalized coordinates can be precisely measured by the encoder for each joint, but the velocity measurements obtained through the tachometers are easily perturbed by noises. Therefore, in order to coincide with these physical constraints, a nonlinear observer proposed in (Bornard G. et al 1993) is used in this paper.
Define the state vector as:
where qi(t) and q̇i(t) are the link position and the velocity of the ith arm respectively. Tεℜ2nx2n is the transformation matrix. The system (3) can be transformed to:
where
With the assumption that the control torque u(t) is uniformly bounded, the high gain observer described in (Bornard G. et al, 1993) can be used to estimate angular positions and angular velocities of the n link rigid robot manipulator (24). The dynamic nonlinear observer is given by:
where K = Γ−1(α) V is the gain of the observer with T(α) = diag(Γi (α)), for any α >0 and due to the observability property of (A,C) the eigenvalues of (A-VC) can be assigned by V.
6. A Simulation example
To illustrate some of the conclusions of this paper, we have simulated the approximate receding-horizon control scheme on a two-link robot arm used in (Lee K. W. & Khalil H. K. 1997), (Spong M. W. 1992) with mechanic equations.
The arm is shown in Figure 1. The dynamic model is described in equation (1) with the following components:
where 1c1 is the mass center of gravity coordinate of the link 1, and 1c2 is the mass center of gravity coordinate of the link 2. The values of the manipulator parameters are given in Table 1 (Lee K. W. & Khalil H. K. 1997) and (Spong M. W. 1992).
Two-linkage manipulators
Physical parameters of two-link arm.
Link 1
m1=10kg l1=1m lc1=0.5m
Link 2
m2=5kg l2 = 1m lc2=0.50m
The reference models chosen in continuous time are:
and .
The nonlinear predictive controller is used to track these desired trajectories with inputs (Lee K. W. & Khalil H. K. 1997):
All simulations are carried with the nonlinear observer (25) with α=0.01 and the assigned eigenvalues σ(λ) = {–0.4, −0.8}. The initial displacements and velocities are chosen as:
The parameter values of two-reference models are chosen as follows: ξ = 1, ω1 = ω2 = 10 rad/s.
The nonlinear controller (9), has been tested by simulation and the control parameters: Q=107In, R=10−14In and h is set to 0.001. Simulation results are show in Figure 2. This Figure gives the angular position (q1(t), q2(t)) and the position tracking error. Although a very short steady state error is observed in the position tracking error and this was expected in the analysis part, a good tracking performance is achieved by this controller in matched case. Figure 3 illustrates the induced control torque applied to robot manipulator. Note that the control torque lie inside the saturation limits (Lee K. W. & Khalil H. K. 1997).
Position tracking error of joint q1 and q2. (matched case).
Induced torque control.
In mismatched case, the frictions are added to the joint of robot manipulator model in equation (1) and are modeled: Fr = fsq + fv sign(q̇), with the values fs = fv = diag(5,5).
If we regard an unknown load carried by the robot as part of the second link, then the parameters m2, lc2 and I2 will change to: m2 + Δm2, lc2 + Δlc2 and I2 + ΔI2.
Let Δm2 = 5Kg, Δlc2 = 0.5m and and to be the maximum parameters variations of the second link due to unknown load. It is observed from Figure 4 that the output q2(t) tracks tightly the reference trajectory with a steady state error. This results has been expected in the analysis did in section 4, i.e the uncertainties will introduce a steady state error in tracking error. Furthermore, the induced torque control lie outside the saturation limits. Figure 5 shows the results when the control law (21) is applied to robot, it is seen from this figure the error was eliminated and the torque control signal lie inside the saturation limits. These results prove the robustness of the rigid link manipulator under the approximate receding-horizon controller with integral action to payload uncertainties and viscous friction.
Position tracking performance in mismatched case.
Torque and tracking performance with integral action.
In addition, other simulations have been carried out and the following remarks have been observed:
Decreases in performances are obtained when we increase the step control parameter h.
High dynamics of the reference trajectories results in high increase of the control torque signal. To reduce this control torque amplitude, one should increase the predictive time increment h. It should be pointed that over a threshold value of h noted hmax, the performance decrease and instability mechanism will appear. This is due to the Taylor approximation used to derive the predictive controller, which becomes invalid.
7. Conclusion
In this paper, the approximation of receding-horizon controller of rigid link robot manipulator using output feedback via link position measurements were considered. Minimizing a quadratic function of the predicted tracking error and the predicted input over the fixed horizon, by using the Simpson's rule approximation, derives the control law. One of the main advantages of these control schemes is that it does not need to perform an online optimization and asymptotic tracking of the smooth reference signal is guaranteed.
To enhance the robustness property of the nonlinear predictive developed by Ping Lu, we proposed to incorporate an integral action in the loop. Simulation shown that payload uncertainties and friction have no effect on the robot manipulator closed by the proposed algorithm. Moreover, the obtained torque signal lie in saturations limit. The Lyapunov theory is used to prove the asymptotic stability of equilibrium point of both original and augmented system.
Finally, we expect that the results presented here can be explored and extended to discrete implementation of these continuous-time predictive controllers either through computers or by special purpose chips that can run at a higher speed.
References
1.
AbdullahC.DawsanP.DoratoP.JamshidiM. (1991), Survey of robust control for rigid robots, IEEE Control Magazine11, pp.24–30.
2.
AtkinsonK. E., (1978) An introduction to numerical analysis, John Wiley, New York.
3.
BornardG.Celle-CouenneF.GillesG., (1993), Observabilite et observateurs, A. J Fossard and D. Normad-Cyrot, Systèmes nonlinéaires: Modélisation et estimation5, Paris, Masson, pp.177–221.
4.
de WitC. CanudasFixotN. (1992), Adaptive control of robot manipulator via velocity estimated feedback, IEEE Transactions On Automatic Control, vol. 37, pp.1234–1237.
5.
de WitC. CanudasFixotN.AstromK. J., (1992), Trajectory tracking in robot manipulators via nonlinear estimated output feedback, IEEE Transaction On Robotic and Automation8, pp.138–144.
6.
ChenW. H.BalanceD. J.GawthropP. J, (2003), Optimal control of nonlinear systems: A predictive approach, Automatica, vol. 39, pp.633–641.
7.
ChenB. S.LeeT. S.FengJ. H. (1994), A nonlinear H∝ control design in robotic systems under parameter perturbation and external disturbances, International Journal of Control, 59(2), pp.439–461.
8.
LeeK. W.KhalilH.K. (1997), Adaptive output feedback control of robot manipulators using high gain observer, International Journal of Control, vol.67, No.6, pp.869–886.
9.
MichalskaH.MayneD. Q. (1993), Robust receding horizon control of constrained nonlinear systems, IEEE Transaction On Automatic Control, 38(11), pp.1623–1633.
10.
NicosiaS.TomeiP., (1990), Robot control by using only joint position measurements, IEEE Transaction On Automatic Control. 35(9), pp.1058–1061.
11.
OrtegaR.SpongM. W. (1989), Adaptive motion control of rigid robots: A tutorial, Automatica25, pp.877–888.
12.
PingL., (1995) Optimal predictive control of continuous nonlinear systems, International Journal of Control, 62(2), pp. 633–649.
13.
PingL, (1998), Approximate nonlinear receding-horizon control laws, International Journal of Control, 71(1), pp.19–34.
14.
SageH. G.de MathelinM. F.OstertagE., (1999), Robust control of robot manipulators: A survey, International Journal of Control, 72(16), 498–522.
15.
SamsonC., (1983), Commande non linéaire robuste des robots manipulateurs, Rapport de recherche No.182, INRIA, France.
16.
SinghS. M. (1995), Nonlinear predictive control of feedback linearizable systems and flight control system design, Journal of guidance, control and dynamics, 18(5), pp.1023–1028.
17.
SlotineJ.J E.SastryS. S., (1983), Tracking control of nonlinear systems using sliding surfaces with application to robot manipulators, International Journal of Control, 38(2), pp.465–492.
18.
SouroukhM.KravarisC., (1996), A continuous-time formulation of nonlinear model predictive control, International Journal of Control, 63(1), pp. 121–146.
19.
SpongM. W.VidyasagarM., (1989), Robot dynamics and control, John Wiley, New York.
20.
SpongM. W. (1992) On robust control of robot manipulators, IEEE Transactions on automatic control, vol.37, No.11, pp.1782–1786.