MCS a.pdf

download MCS a.pdf

of 84

Transcript of MCS a.pdf

  • 7/25/2019 MCS a.pdf

    1/84

    1

    NPR COLLEGE OF ENGINEERING &TECHNOLOGY

    BE EEE-III/ SEMESTER VI

    EE1354MODERN CONTROL SYSTEMS

    Prepared By:

    A.R.SALINIDEVI Lect/EEE

  • 7/25/2019 MCS a.pdf

    2/84

    2

    EE1354MODERN CONTROL SYSTEMS(Common to EEE, EIE and ICE)

    L T P C

    3 1 0 4

    UNIT I STATE SPACE ANALYSIS OF CONTINUOUS TIME SYSTEMS 9State variable representationConversion of state variable form to transfer function and vice

    versaEigenvalues and EigenvectorsSolution of state equationControllability and

    observabilityPole placement designDesign of state observer

    UNIT II z-TRANSFORM AND SAMPLED DATA SYSTEMS 9

    Sampled data theorySampling processSampling theoremSignal reconstruction

    Sample and hold circuitsz-TransformTheorems on z-TransformsInverse z-Transforms

    Discrete systems and solution of difference equation using z transform Pulse transfer

    functionResponse of sampled data system to step and ramp InputsStability studiesJurys test and bilinear transformation

    UNIT III STATE SPACE ANALYSIS OF DISCRETE TIME SYSTEMS 9State variablesCanonical formsDigitalizationSolution of state equations

    Controllability and ObservabilityEffect of sampling time on controllabilityPole

    placement by state feedbackLinear observer designFirst order and second order

    problems

    UNIT IV NONLINEAR SYSTEMS 9Types of nonlinearityTypical examplesPhase-plane analysisSingular pointsLimit

    cyclesConstruction of phase trajectoriesDescribing function methodBasic concepts

    Dead ZoneSaturationRelayBacklashLiapunov stability analysisStability in the

    sense of LiapunovDefiniteness of scalar functionsQuadratic formsSecond method ofLiapunovLiapunov stability analysis of linear time invariant systems and non-linear system

    UNIT V MIMO SYSTEMS 9Models of MIMO systemMatrix representationTransfer function representationPoles

    and ZerosDecouplingIntroduction to multivariable Nyquist plot and singular values

    analysisModel predictive control

    L: 45 T: 15 Total: 60

    TEXT BOOKS1. Gopal, M., Digital Control and State Variable Methods, 3rd Edition, Tata McGraw Hill,

    2008.

    2. Gopal, M., Modern Control Engineering, New Age International, 2005.

    REFERENCES1. Richard C. Dorf and Robert H. Bishop, Modern Control Systems, 8th Edition, Pearson

    Education, 2004.

    2. Gopal, M., Control Systems: Principles and Design, 2nd Edition, Tata McGraw Hill,

    2003.

    3. Katsuhiko Ogata, Discrete-Time Control Systems, Pearson Education, 2002.

  • 7/25/2019 MCS a.pdf

    3/84

    3

    MODERN CONTROL SYSTEM

    Unit I

    STATE SPACE ANALYSIS OF CONTINUOUS TIME SYSTEMS

    State Variable Representation

    The state variables may be totally independent of each other, leading

    to diagonal or normal form or they could be derived as the derivatives of the output. If

    them is no direct relationship between various states. We could use a suitable

    transformation to obtain the representation in diagonal form.

    Phase Variable Representation

    It is often convenient to consider the output of the system as one of the state

    variable and remaining state variable as derivatives of this state variable. The state

    variables thus obtained from one of the system variables and its (n-1) derivatives, are

    known as n-dimensional phase variables.

    In a third-order mechanical system,the output may be displacement

    vxxx 2

    .

    1,1 and axx 32.

    in the case of motion of translation or angular displacement

    wxxx 2

    .

    1,11 and.

    3

    .

    2 wxx if the motion is rotational,

    Where v ,,, awv respectively, are velocity, angular velocity acceleration, angular

    acceleration.

    Consider a SISO system described by nth-order differential equation

    Where

    u is, in general, a function of time.

    The nth order transfer function of this system is

    State variable representation Conversion of state variable form to transfer function and

    vice versa Eigenvalues and Eigenvectors Solution of state equation Controllability

    and observabilityPole placement designDesign of state observer

  • 7/25/2019 MCS a.pdf

    4/84

    4

    With the states (each being function of time) be defined as

    Equation becomes

    Using above Eqs state equations in phase satiable loan can he obtained as

    Where

    Physical Variable Representation

    In this representation the state variables are real physical variables, which can

    be measured and used for manipulation or for control purposes. The approach generally

    adopted is to break the block diagram of the transfer function into subsystems in such a

    way that the physical variables can he identified. The governing equations for the

    subsystems can he used to identify the physical variables. To illustrate the approach

    consider the block diagram of Fig.

    One may represent the transfer function of this system as

    Taking H(s) = 1, the block diagram of can be redrawn as in Fig. physical variables can

    be speculated as x1=y, output,.

    2 wx the angular velocity aix3 the armature

    current in a position-control system.

  • 7/25/2019 MCS a.pdf

    5/84

    5

    Where

    The state space representation can be obtained by

    And

    State space models from transfer functions

    A simple example of system has an input and output as shown in Figure 1. This class

    of system has general form of model given in Eq.(1).

    1 1

    1 0 1 01 1( ) ( )

    n n m

    n mn n m

    d y d y d ua a y t b b u t

    dt dt dt

    Models of this form have the property of the following:

    1 1 2 2 1 1 2 2( ) ( ) ( ) y( ) ( ) ( )u t u t u t t y t y t (2)

    where, (y1, u1)and (y2,u2)each satisfies Eq,(1).

    Model of the form of Eq.(1) is known as linear time invariant (abbr. LTI) system. Assume

    the system is at rest prior to the time t0=0, and, the input u(t) (0 t < )produces the output

    y(t) (0 t < ), the model of Eq.(1) can be represented by a transfer function in term of

    Laplace transform variables, i.e.:

    Sy(t)u(t)

  • 7/25/2019 MCS a.pdf

    6/84

    6

    1

    1 0

    1

    1 0

    ( ) ( )m m

    m m

    n n

    n n

    b s b s by s u s

    a s a s a

    (3)

    Then applying the same input shifted by any amount of time produces the same output

    shifted by the same amount q of time. The representation of this fact is given by the following

    transfer function:

    1

    1 0

    1

    1 0

    ( ) ( )m m

    sm m

    n n

    n n

    b s b s by s e u s

    a s a s a

    (4)

    Models of Eq.(1) having all 0 ( 0)ib i , a state space description arose out of a reduction

    to a system of first order differential equations. This technique is quite general. First, Eq.(1)

    is written as:

    ( ) ( 1)

    ( 1)

    0 1 1

    , ( ), , . , , ;

    with initial conditions: y(0)=y , (0) (0), , (0) (0)

    n n

    n

    n

    y f t u t y y y y

    y y y y

    (5)

    Consider the vector nx R with ( 1)1 2 3, , , , n

    nx y x y x y x y , Eq.(5) becomes:

    2

    3

    ( 1), ( ), , . , ,

    n

    n

    x

    xd

    Xdt

    x

    f t u t y y y y

    (6)

    In case of linear system, Eq.(6) becomes:

    0 1 n-1

    0 1 0 0 0

    0 0 1 0 0 0

    0 ( ); y(t)= 1 0 0 0

    0 0 1

    -a -a -a 1

    dX X u t X

    dt

    (7)

    It can be shown that the general form of Eq.(1) can be written as

    0 1 m

    0 1 n-1

    0 1 0 0 0

    0 0 1 0 0 0

    0 ( ); y(t)= b b b 0 0

    0 0 1

    -a -a -a 1

    dX X u t X

    dt

    (8)

    and, will be represented in an abbreviation form:

  • 7/25/2019 MCS a.pdf

    7/84

    7

    ; ; D=X AX Bu y CX Du 0 (9)

    Eq.(9) is known as the controller canonical form of the system.

    Transfer function from state space modelsWe have just showed that a transfer function model can be expressed as a state space

    system of controller canonical form. In the reverse direction, it also easy to see that each

    linear state space system of Eq.(9) cab be expressed as a LTI transfer function. The procedure

    is to take laplace transformation of the both sides of Eq,(9) to give:

    ( ) ( ) ( ) ; ( ) ( ) ( )sX s AX s Bu s y s CX s Du s (10)

    So that

    1

    ( )( ) ( ) ( ) ( )( )

    n sy s C sI A B D u s G s u sd s

    (11)

    An algorithm to compute the transfer function from state space matrices is given by the

    Leverrier-Fadeeva-Frame formula of the following:

    1

    1 2

    0 1 2 1

    1

    1 1

    0 1 0

    1 0 1 2 1

    ( )

    ( )

    ( )

    ( )

    ,

    ( )

    1/ 2 ( )

    n n

    n n

    n n

    n n

    N ssI A

    d s

    N s s N s N sN N

    d s s d s d s d

    where

    N I d trace AN

    N AN d I d trace AN

    1 2 1 1 2

    -1 1

    1 ( )

    1

    1 0 A

    n n n n n

    n n n n

    N AN d I d trace ANn

    N d I d trace ARn

    (12)

    Therefore, according to the algorithm mentioned, the transfer function becomes:

    ( ) ( )n s CN s B CD (13)

    or,( )

    ( )( )

    CN s B CDG s

    d s (14)

    EigenValues

    Consider an equation AX = Y which indicates the transformation of 1n vector

    matrix X into 'n x 1' vector matrix Y by 'n x n' matrix operator A.

  • 7/25/2019 MCS a.pdf

    8/84

    8

    If there exists such a vector X such that A transforms it to a vector XX then X is

    called the solution of the equation,

    The set of homogeneous equations (1) have a nontrivial solution only under

    thecondition,

    The determinant | X I - A | is called characteristic polynomial while the equation (2)

    is called the characteristic equation.

    After expanding, we get the characteristic equation as,

    The 'n' roots of the equation (3) i.e. the values of X satisfying the above equation

    (3) are called eigen values of the matrix A.

    The equation (2) is similar to| sI- A | =0, which is the characteristic equation of the

    system. Hence values of X satisfying characteristic equation arc the closed loop

    poles ofthe system. Thus eigen values are the closed loop poles of the system.

    Eigen Vectors

    Any nonzero vector iX such that iii XAX is said to be eigen vector associated

    with eigenvalue i .Thus let i satisfies the equation

    Then solution of this equation is called eigen vector of A associated with eigen

    value i and is denoted as Mi.

    If the rank of the matrix [ i I - A] is r, then there are (n - r) independent Eigen

    vectors. Similarly another important point is that if the eigenvalues of matrix A

    are all distinct, then the rank r of matrix A is (n - 1) where n is order of the

    system.

  • 7/25/2019 MCS a.pdf

    9/84

    9

    Mathematically, the Eigen vector can be calculated by taking cofactors of

    matrix ( i I - A) along any row.

    Where C ki is cofactor of matrix ( i I - A) of kthrow.

    Key Point: If the cofactor along a particular row gives null solution i.e. all elements of

    corresponding eigen vectors are zero then cofactors along any other row must he

    obtained. Otherwise inverse of modal matrix M cannot exist.

    Example 1

    Obtain the Eigen values, Eigen vectors for the matrix

    Solution

    Eigen values are roots of

    Eigen values are

    To find Eigen vector,

    Let

  • 7/25/2019 MCS a.pdf

    10/84

    10

    Where C = cofactor

    For 2 = -2

    For 3 = -3

    Example 2

    For a system with state model matrices

    Obtain the system with state model matrices

  • 7/25/2019 MCS a.pdf

    11/84

    11

    Solution

    The T.F. is given by,

  • 7/25/2019 MCS a.pdf

    12/84

    12

    Solution of State Equations

    Consider the state equation n of linear time invariant system as,

    )()()(.

    tBUtAXtX

    The matrices A and B are constant matrices. This state equation can be of two types,

    1. Homogeneous and

    2. Nonhomogeneous

    Homogeneous Equation

    If A is a constant matrix and input control forces are zero then the equation

    takes the form,

    Such an equation is called homogeneous equation. The obvious equation is if input is

    zero, In such systems, the driving force is provided by the initial conditions of the

    system to produce the output. For example, consider a series RC circuit in which

    capacitor is initially charged to V volts. The current is the output. Now there is no

    input control force i.e. external voltage applied to the system. But the initial voltage on

    the capacitor drives the current through the system and capacitor starts

    discharging through the resistance R. Such a system which works on the initial

    conditions without any input applied to it is called homogeneous system.

    Nonhomogeneous Equation

    If A is a constant matrix and matrix U(t) is non-zero vector i.e. the input

    control forces are applied to the system then the equation takes normal form as,

    Such an equation is called nonhomogeneous equation. Most of the practical

    systems require inputs to dive them. Such systems arc nonhomogeneous linear

    systems.

    The solution of the state equation is obtained by considering basic method of

  • 7/25/2019 MCS a.pdf

    13/84

    13

    finding the solution of homogeneous equation.

    Controllability and Observability

    More specially, for system of Eq.(1), there exists a similar transformation that will

    diagonalize the system. In other words, There is a transformation matrix Q such that

    0 ; ; X(0)=XX AX Bu y CX Du (1)

    -1 or X=QX QX X

    (2)

    y = CX X Bu X Du

    (3)

    Where

    1

    2

    0 0

    0 0

    0 n

    (4)

    Notice that by doing the diagonalizing transformation, the resulting transfer function between

    u(s) and y(s) will not be altered.

    Looking at Eq.(3), if 0kb

    , then kx (t) is uncontrollable by the input u(t), since, kx (t) is

    characterized by the mode kte by the equation:

    ( ) (0 )kt

    k kx t e x

    The lake of controllability of the state kx (t) is reflect by a zero kthrow ofB

    , i.e.

    kb

    . Which

    would cause a complete zero rows in the following matrix (known as the controllability

    matrix), i.e.:

    C(A,b)

    2 1

    1 1 1 1 1 1 1

    2 1

    2 2 2 2 2 2 2

    2 3 n-1

    2 1

    k k k

    A A A A

    n

    n

    n

    k k k k

    b b b b

    b b b b

    B B B B Bb b b b

    2 1

    n n n

    n

    n n n nb b b b

    (5)

    A C(A,b) matrix with all non-zero row has a rank of N.

    In fact , 1 orB Q B B QB

    . Thus, a non-singular C(A,b) matrix implies a non-singular

    matrix of C(A,b)of the following:

  • 7/25/2019 MCS a.pdf

    14/84

    14

    C(A,b) 2 -1 nB AB A B A B (6)

    It is important to note that this result holds in the case of non-distinct eigenvalues as well.

    Remark 1]

    If matrix A has distinct eigenvalues and is represented as a controller canonical form, it

    is easy to show the following identity holds, i.e.:

    2 1 2 1

    1 1 1 1 1 1 11 1n nA f or each i.

    Therefore a transpose of so-called Vandermonde matrix V of n column eigenvectors of A will

    diagonalize A, i.e.,

    2 1

    1 1 11 2 2 12 2 2 2 2 2

    1 2

    2 n-1

    n1 1 n-1

    1 2 n

    1 1 11

    1

    1

    T

    n

    n n

    Tn

    n nn n

    W

    (6)

    and

    1 1T T

    or, A= AT T T T W A W W W W W A

    [Remark 2]

    There is an alternative way to explain why C(A,b) should have rank n for state controllable,

    let us start from the solution of the state space system:

    0

    ( )( ) (0 ) ( )

    ft

    At A t

    t

    X t e X e Bu d (7)

    The state controllability requires that for each X(tf) nearby X(t0), there is a finite sequence of

    u(t; t [to,tf]).

  • 7/25/2019 MCS a.pdf

    15/84

    15

    0

    0

    0

    0

    0

    0

    0

    0

    ( 1)

    0

    0

    ( 1) 1

    0

    0

    ( ) ( )

    ( ) ( )

    ( )

    = ( ) ( )

    f

    f

    f

    f

    t

    At A

    f

    t

    t

    AtA

    f

    t

    t knA

    k t k

    t k i ni

    i

    k it k

    X t e X e Bu d

    or

    e Bu d e X t X

    e Bu t k d

    A B u t k

    0

    0

    0

    ( 1)i=n-1

    0i=0 0

    1

    i=n-122 n-1

    i=0

    = ( ) ( )

    = AB A B A B

    n

    t k ki

    ikt k

    i

    i

    n

    d

    A B u t k d

    w

    wA BW B

    w

    Thus, in order W has non-trival solution, we need that C(A,b) matrix has exact rank n

    There are several alternative ways of establishing the state space controllability:

    The (n) rows of Ate B are linearly independent over the real field for all t.

    The controllability grammian

    0

    ( , )

    fT

    t

    At T A

    ram o f

    t

    G t t e BB e d is non-singular for all 0ft t .

    [Theorem 1]Replace B with b, (i.e. Dim{B}=n 1), a pair [A,b] is non-controllable if

    and only if there exists a row vector 0q such that

    , 0qA q qb (8)

    To prove the if part:

    If there is such row vector, we have:

  • 7/25/2019 MCS a.pdf

    16/84

    16

    2 2 2 1

    -1 1

    0

    0 0

    0 , , , , 0

    and 0

    0

    n

    n n

    qA q and qb

    qAb qbq I A

    qA b qAb qb q b Ab A b A b

    qb

    qA b qb

    Since 0q , we conclude that: 2 1, , , , nb Ab A b A b is singular, and thus the

    system is not controllable.

    To prove the only if part:

    If the pair is noncontrollable, then matrix A can be transformed into non-controllable form like:

    ,0 0

    C CCC

    C

    A A rbA b

    A n r (9)

    Where, r rank C(A,b) (Notice that Eq.(33) is a well-known theorem in linear

    system.)

    Thus, one can find a row vector has the form [0 ]q z , where z can be selected as

    the eigenvector of CA , (i.e.: CzA z), for then:

    0 z 0qA A z q (10)

    Therefore, we have shown that only if [A, b] is non-controllable, there is a

    non-zero row vector satisfying Eq.(8).

    In fact, according to Eq.(27),

    1

    1

    i

    ktAt t t T T

    i i

    i

    e Ve V Ve W v w e

    and, 0( ) AtX t e X , we have:

    ( )( )

    0 0

    1 10 0

    ( ) ( ) ( )i iit tk n

    t tAt A t T T

    i i i i

    i i

    X t e X e bu d v w e X v w b e u d

  • 7/25/2019 MCS a.pdf

    17/84

    17

    Thus, if b is orthogonal to iw , then the state associated with i will not be controllable,

    and hence, the system is not completely controllable. The another form to test for the

    controllability of the [A,b] pair is known as the Popov-Belevitch-Hautus (abbrv. PBH) test is

    to check if rank sI A b n for all s (not only at eigenvalues of A). This test is based on

    the fact that if sI A b has rank n, there cannot be a nonzero row vector q satisfying

    Eq.(32). Thus by Theorem 1, pair [A, b] must be controllable.

    Referring to the systems described by Eqs.(26) and (27), the state ( )ix t

    corresponding to

    the mode ite is unobservable at the output 1y , if 1 0iC

    for any i=1,2,,n. The lack of

    observability of the state ( )ix t

    is reflected by a complete zero (ith) column of so called

    observability matrix of the system O ( , )A C

    , i.e.:

    O1( , )A C

    11 12 11

    1 11 2 12 11

    1 1 11

    1 2 12 11

    n

    n n

    n n nn

    n n

    C C CC

    C C CC A

    C C CC A

    (11)

    An observable state ( )ix t corresponds to a nonzero column of O ( , )A C . In the case of distinct

    eigenvalues, each nonzero column increases the rank by one. Therefore, the rank of

    O ( , )A C

    corresponding to the total number of modes that are observable at the output y(t) is

    termed the observability rank of the system. As in the case of controllability, it is not

    necessary to transform a given state-space system to modal canonical form in order to

    determine its rank. In general, the observability matrix of the system is defined as:

    O ( , )A C =

    1n

    C

    CA

    CA

    = O

    1( , ) ( , )A C Q A C V

    With Q=V-1nonsingular. There, the rank of O ( , )A C equals the rank of O ( , )A C

    . It is

    important to note that this result holds in the case of non-distinct eigenvalues. Thus, a state-

    space system is said to be completely (state) observable if its observability matrix has a full

    rank n. Otherwise the system is said to be unobservable

  • 7/25/2019 MCS a.pdf

    18/84

    18

    In particular, it is well known that a state-space system is observable if and only if the

    following conditions are satisfied:

    The (n) column ofAt

    Ce are linearly independent over R for all t.

    The observability grammian of the following is nonsingular for all 0ft t :

    0

    ,

    T

    t

    A T A

    ranm o

    t

    G e C Ce d

    The (n+p) n matrixI A

    Cb has rank n at all eigenvalues i of A.

    Pole Placement Design

    The conventional method of design of single input single output control system

    consists of design of a suitable controller or compensator in such a way that the dominant

    closed loop poles will have a desired damping ratio % and undamped natural frequency con.

    The order of the system in this case is increased by 1 or 2 if there are no pole zero

    cancellation taking place. It is assumed in this method that the effects on the responses of non-

    dominant closed loop poles lo be negligible. Instead of specifying only the dominant closed

    loop poles in the conventional method of design, the pole placement technique describes all the

    closed loop poles which require measurements of all state variables or inclusion of a state

    observer in the system. The system closed loop poles can be placed at arbitrarily chosen

    locations with the condition that the system is completely stale controllable. This condition

    can be proved and the proof is given below. Consider a control system described by following

    slate equation

    Here x is a state vector, u is a control signal which is scalar, A is n x n state matrix. B is n x 1

    constant matrix.

    Fig open loop control system

  • 7/25/2019 MCS a.pdf

    19/84

    19

    The system defined by above equation represents open loop system. The

    state x is not fed back to the control signal u. Let us select the control signal to

    be u = - Kx state. This indicates that the control signal is obtained from instantaneous state.

    This is called state feedback. The k is a matrix of order l x n called state feedback gain matrix.

    Let us consider the control signal to be unconstrained. Substituting value of u in equation 1

    The system defined by above equation is shown in the Fig. 5.2. It is a closed loop

    control system as the system state x is fed back to the control system as the system stale x

    is fed back to control signal u. Thus this a system with state feedback

    The solution of equation 2 is say

    x(t) = e,x(0) is the initial slate (3)

    The stability and the transient response characteristics are determined by the eigen

    values of matrix A - BK. Depending on the selection of state feedback gain matrix K, the

    matrix A - BK can be made asymptotically stable and it is possible to make x(t) approaching

    to zero as time t approaches to infinity provided x(0) * 0. The eigen values of matrix A - BK

    arc called regulator poles. These regulator poles when placed in left half of s plane then x(t)

    approaches zero as time t approaches infinity. The problem of placing the closed loop poles

    at the desired location is called a poleplacement problem.

    Design of State Observer

    In case of state observer, the state variables are estimated based on the

    measurements of the output and control variables. The concept of observability

    plays important part here in case of state observer.

    Consider a system defined by following state equations

  • 7/25/2019 MCS a.pdf

    20/84

    20

    Let us consider x as the observed state vectors. The observer is basically a

    subsystem which reconstructs the state vector of the system. The mathematical

    model of the observer is same as that of the plant except the inclusion of additional

    term consisting of estimation error to compensate for inaccuracies in matrices A and

    B and the lack of the initial error.

    The estimation error or the observation error is the difference between the

    measured output and the estimated output. The initial error is the difference

    between the initial state and the initial estimated state. Thus the mathematical

    model of the observer can be defined as,

    Here x is the estimated state and C x is the estimated output. The observer has

    inputs of output y and control input u. Matrix K^ is called the observer gain

    matrix. It is nothingbut weighing matrix for the correction term which contains

    the difference between the measured output y and the estimated output cx

    This additional term continuously corrects the model output and the performance

    of the observer is improved.

    Full order state observer

    The system equations arc already defined as

    The mathematical model of the state observer is taken as.

    To determine the observer error equation, subtracting equation of x from x wc get

  • 7/25/2019 MCS a.pdf

    21/84

    21

    The block diagram of the system and full order state observer is shown in the Fig.

    The dynamic behavior of the error vector is obtained from the Eigen values of matrix

    A-K^C If matrix A-K^C is a stable matrix then the error vector will converge to

    zero for any initial error vector e(0). Hence x(t) will converge to x(t) irrespective of

    values of x(0) and x(0).

    If the Eigen values of matrix A-KeC are selected in such a manner that the dynamic

    behavior of the error vector is asymptotically stable and is sufficiently fast then

    any of the error vector will tend to zero with sufficient speed.

    1/ the system is completely observable then it can be shown that it is possible to select

    matrix K,. such that A-K^C has arbitrarily desired Eigen values, i.c. observer gain

    matrix Ke can be obtained to get the desired matrix A-KCC.

  • 7/25/2019 MCS a.pdf

    22/84

    22

    UNIT I

    STATE SPACE ANALYSIS OF CONTINUOUS TIME SYSTEMS

    PART A

    1.

    What are the advantages of state space analysis?

    2.

    What are the drawbacks in transfer function model analysis?

    3.What is state and state variable?

    4.What is a state vector?

    5.Write the state model of nthorder system?

    6.What is state space

    7.What are phase variables?

    8.

    Write the solution of homogeneous state equation?

    9.Write the solution of nonhomogeneous state equation?

    10. What is resolvant matrix?

    PART B

    1.Explain Kamans test for determining state controllability?

    2.Explain Gilberts test for determining state controllability?

    3.Find the output of the system having state model,

    and

    The input U(t) is unit step and X(0)0

    10

    4.Show the following system is completely state controllable and observable.

    And

    5.

    Obtain the homogenous solution of the equation X(t) =A X(t)

    6.

    Derive the transfer function of observer based controller?

  • 7/25/2019 MCS a.pdf

    23/84

    23

    UNIT II

    Z-TRANSFORM AND SAMPLED DATA SYSTEMS

    Sampled Data System

    When the signal or information at any or some points in a system is in the form of

    discrete pulses. Then the system is called discrete data system. In control engineering the

    discrete data system is popularly known as sampled data systems.

    Sampling process

    Sampling is the conversion of a continuous time signal into a discrete time signal

    obtained by taking sample of the continuous time signal at discrete time instants.

    Thus if f (t) is the input to the sampler

    The output is f(kT)

    Where T is called the sampling interval

    The reciprocal of T

    Sampled data theory Sampling process Sampling theorem Signal reconstruction Sample and hold circuits z-Transform Theorems on z- Transforms Inverse z-

    Transforms Discrete systems and solution of difference equation using z transform

    Pulse transfer function Response of sampled data system to step and ramp Inputs

    Stability studiesJurys test and bilinear transformation

  • 7/25/2019 MCS a.pdf

    24/84

    24

    Let 1/T =Fsis called the sampling rate. This type of sampling is called periodic

    Sampling, since samples are obtained uniformly at intervals of T seconds.

    Multiple order samplingA particular sampling pattern is repeated periodically

    Multiple rate sampling - In this method two simultaneous sampling operations with

    different time periods are carried out on the signal to produce the sampled output.

    Random samplingIn this case the sampling instants are random

    Sampling Theorem

    A band limited continuous time signal with highest frequency fmhertz can be uniquely

    recovered from its samples provided that the sampling rate Fsis greater than or equal to 2fm

    samples per seconds

    Signal Reconstruction

    The signal given to the digital controller is a sampled data signal and in turn the

    controller gives the controller output in digital form. But the system to be controlled needs an

  • 7/25/2019 MCS a.pdf

    25/84

    25

    analog control signal as input. Therefore the digital output of controllers must be converters

    into analog form

    This can be achieved by means of various types of hold circuits. The simplest holdcircuits are the zero order hold (ZOH). In ZOH, the reconstructed analog signal acquires the

    same values as the last received sample for the entire sampling period

    The high frequency noises present in the reconstructed signal are automatically

    filtered out by the control system component which behaves like low pass filters. In a first

    order hold the last two signals for the current sampling period. Similarly higher order hold

    circuit can be devised. First or higher order hold circuits offer no particular advantage over

    the zero order hold

    Z- Transform

    Definition of Z Transform

    Let f (k) = Discrete time signal

    F (z) = Z {f (k)} =z transform of f (k)

    The z transforms of a discrete time signal or sequence is defined as the power series

    k

    k

    zkfZF )()( -------------- 1

    Where z is a complex variable

    Equation (1) is considered to be two sided and the transform is called two sided z transform.

    Since the time index k is defined for both positive and negative values.

    The one sided z transform of f(k) is defined as

    k

    k

    zkfZF

    0

    )()( --------------- 2

  • 7/25/2019 MCS a.pdf

    26/84

    26

    Problem

    1. Determine the z transform and their ROC of the discrete sequences f(k) ={3,2,5,7}

    Given f (k) = {3, 2, 5, 7}

    Where f (0) =3

    f (1) =2

    f (2) = 5

    f (3) = 7

    f (k) = 0 for k < 0 and k >3

    By definition

    k

    k

    zkfzFkfZ )()()}({

    The given sequence is a finite duration sequence. Hence the limits of summation can be

    changed as k = 0 to k = 3

    k

    k

    zkfzF3

    0

    )()(

    0)0()( zfzF + 1)1( zf + 2)2( zf + 3)3( zf

    = 321 7523 zzz

    Here )(zF is bounded, expect when z =0

    The ROC is entire z-plane expect z = 0

    2. Determine the z transform of discrete sequences f (k) =u (k)

    Given f (k) =u (k)

    u (k) is a discrete unit step sequence

    u (k) = 1 for k 0

    = 0for k < 0

    By definition

    k

    k

    zkfzFkfZ )()()}({ k

    k

    zku0

    )(

    k

    k

    z0

    k

    k

    z )( 1

    0

  • 7/25/2019 MCS a.pdf

    27/84

    27

    F (z) is an infinite geometric series and it converges if 1z

    F (z)11

    1

    z

    z

    11

    1

    1z

    z

    3. Find the one sided z transform of the discrete sequences generated by mathematically

    sampling the continuous time function f (t) kTe at cos

    Given

    f (t) kTe at cos

    By definition

    F(z) = k

    k

    akT zkTekfZ0

    cos)}({

    k

    k

    TkjTkjTka z

    eee

    0 2

    0 0

    11

    2

    1

    2

    1

    k k

    kTjTakTjTa zeezee

    WKTc

    ck

    k

    1

    1

    0

    11 1

    1

    2

    1

    1

    1

    2

    1)(

    zeezeezF

    TwjTaTwjaT

    aT

    Tj

    aT

    Tj

    eze

    eze 1

    1

    2

    1

    1

    1

    2

    1

    TjaT

    aT

    TjaT

    aT

    eez

    ez

    eez

    ez

    2

    1

    TjaTTjaT

    TjaTaTTjaTaT

    ezeeze

    ezezeezeze

    2

    1

    TjTjTjaTTjaTaT

    TjaTTjaTaT

    eeezeezeze

    ezeezeze2)(2

  • 7/25/2019 MCS a.pdf

    28/84

    28

    1)(

    )(2

    2 22 TjTjaTaT

    TjTjaTaT

    eezeez

    eezeze

    1cos2

    )cos(

    2 22 Tzeez

    TzezezeaTaT

    aTaTaT

    2

    cosjj ee

    Inverse z transform

    Partial fraction expansion (PFE)

    Power series expansion

    Partial fraction expansion

    Let f (k) =discrete sequence

    F (z) =Z {f (k)} = z transform of f (k)

    F (z) =n

    nnn

    m

    mmm

    azazaz

    bzbzbzb

    ..........

    ..........2

    2

    1

    1

    2

    0

    1

    00 where nm

    The function F (z) can be expressed as a series of sum terms by PFE

    n

    i i

    i

    pz

    AAzF

    1

    0)( -------------- 3

    Where0

    A is a constant

    nAAA ,........, 21 are residues

    nppp ,........, 21 are poles

    Power series expansion

    Let f (k) =discrete sequences

    F (z) = Z {f (k)} = z transform of f (k)

    By definition

    k

    k

    zkfZF )()(

    On Expanding

    ...)..........)2()1()0()1()2()3((.......)( 210123 zfzfzfzfzfzfZF -------4

    Problem

    1. Determine the inverse z transform of the following function

    (i)F (z) =21 5.05.11

    1

    zz

  • 7/25/2019 MCS a.pdf

    29/84

    29

    Given

    F (z) =21

    5.05.11

    1

    zz

    2

    5.05.11

    1

    zz

    5.05.12

    2

    zz

    z

    )5.0()1(

    2

    zz

    z

    )5.0()1(

    )(

    zz

    z

    z

    zF

    By partial fraction expansion

    )5.0()1(

    )( 21

    z

    A

    z

    A

    z

    zF

    1A )1()(

    zz

    zF

    Put z =1

    1A )1()5.0()1(

    zzz

    z

    )5.0(z

    z

    )5.01(

    1

    2

    Put z =0.5

    2A )5.0()(

    zz

    zF

    2A )5.0()5.0()1(

    zzz

    z

    )1(z

    z

  • 7/25/2019 MCS a.pdf

    30/84

    30

    )15.0(

    5.0

    -1

    5.0

    1

    1

    2)(

    zzz

    zF

    5.01

    2)(

    z

    z

    z

    zzF

    WKT

    az

    zaZ k}{ and

    1)}({

    z

    zkuZ

    On taking inverse z transform

    0,)5.0()(2)( kkukf k

    (ii)F (z) =5.0

    2

    2

    zz

    z

    Given

    F (z) =5.0

    2

    2

    zz

    z

    )5.05.0()5.05.0(

    2

    jzjz

    z

    )5.05.0()5.05.0(

    )(

    jzjz

    z

    z

    zF

    By partial fraction expansion

    )5.05.0()5.05.0(

    )( *

    jz

    A

    jz

    A

    z

    zF

    A )5.05.0()(

    jzz

    zF

    Put z = 0.5+j0.5

    A )5.05.0()5.05.0()5.05.0(

    jzjzjz

    z

    )5.05.0( jz

    z

  • 7/25/2019 MCS a.pdf

    31/84

    31

    )5.05.05.05.0(

    5.05.0

    jj

    j

    5.05.0 j

    *A )5.05.0(

    )5.05.0()5.05.0(jz

    jzjz

    z

    Put z =0.5-j0.5

    *A

    )5.05.0( jz

    z

    )5.05.05.05.0(

    5.05.0

    jj

    j

    5.05.0 j

    )5.05.0()5.05.0(

    )5.05.0()5.05.0()(

    jzj

    jzj

    zzF

    )5.05.0()5.05.0(

    )5.05.0()5.05.0(

    jzzj

    jzzj

    WKTaz

    zaZ k}{

    On taking inverse z transform

    kk jjjjkf )5.05.0()5.05.0()5.05.0)(5.05.0()(

    kk j

    j

    jj

    j

    j )5.05.0()5.05.0

    ()5.05.0()5.05.0

    (

    kk jjjjjj )5.05.0()5.05.0()5.05.0)(5.05.0(

    11 )5.05.0()5.05.0( kk jjjj

    2. Determine the inverse z transform of z domain function

    F (z) =23

    1232

    2

    zz

    zz

    Given

    F (z) =23

    1232

    2

    zz

    zz

    3

    232 zz 123 2 zz

    693 2 zz

    511z

  • 7/25/2019 MCS a.pdf

    32/84

    32

    F (z) =23

    5113

    2 zz

    z

    )2()1(

    5113

    zz

    z

    By PFE

    F (z) =)2()1(

    3 21

    z

    A

    z

    A

    1A )1()2()1(

    511z

    zz

    z

    )2(

    511

    z

    z6

    21

    511

    2A )2()2()1(

    511z

    zz

    z

    )1(

    511

    z

    z

    1712

    5)2(11

    )2(

    17

    )1(

    63)(

    zzzF

    2

    117

    )1(

    163

    z

    z

    zz

    z

    z

    217

    )1(63

    11

    z

    zz

    z

    zz

    On taking inverse z transform

    0;)1(217)1(6)(3)( )1( kforkukukkf k

    2. Determine the inverse z transform of the following

    21

    2

    1

    2

    3

    1

    1)(

    zz

    zF

    Where (i) ROC 0.1z

    (ii) ROC 5.0z

    Given

    21

    2

    1

    2

    31

    1)(

    zz

    zF

    (i)

    0.1z

  • 7/25/2019 MCS a.pdf

    33/84

    33

    .........8

    15

    4

    7

    2

    31 321 zzz

    21

    2

    1

    2

    31 zz 1

    21

    2

    1

    2

    31 zz

    321

    21

    4

    3

    4

    9

    2

    3

    2

    1

    2

    3

    zzz

    zz

    422

    22

    8

    7

    4

    3

    4

    7

    43

    47

    zzz

    zz

    43

    8

    7

    8

    15zz

    )(zF .........8

    15

    4

    7

    2

    31

    321 zzz ------------ -(i)

    k

    k

    zkfZF )()(

    For a causal signal

    k

    k

    zkfzF0

    )()(

    .................)2()1()0()( 321 zfzfzfzF

    --------------(ii)

    Comparing equation (i) &(ii)

    1)0(f ,2

    3)1(f ,

    4

    7)2(f ,

    8

    15)3(f

    0......}..........,8

    15,

    4

    7,

    2

    3,1{)( kforkf

  • 7/25/2019 MCS a.pdf

    34/84

    34

    (i)

    5.0z

    ..........301162 5432 zzzz

    12

    3

    2

    1 12 zz 1

    2231 zz

    32

    2

    693

    23

    zzz

    zz

    432

    2

    14217

    367

    zzzzz

    543

    43

    304515

    1415

    zzz

    zz

    )(zF ..........301162 5432 zzzz -------------- (i)

    k

    k

    zkfZF )()(

    For an anti-causal signal

    10

    )()( zkfZFk

    )0()1()2()3()4()5(............)( 12345 fzfzfzfzfzfzF ------------- (ii)

    Comparing the equation i & ii

    30)5(f , 14)4(f , 6)3(f , 2)2(f , 0)1(f , 0)0(f

    }0,0,2,6,14,30..{.........)(kf

    Difference equation

    Discrete time systems are described by difference equation of the form

  • 7/25/2019 MCS a.pdf

    35/84

    35

    If the system is causal, a linear difference equation provides an explicit relationship between

    the input and output. This can be seen by rewriting.

    Thus the nth value of the output can be computed from the nth input value and the N and M

    past values of the output and input, respectively.

    Role of z transform in linear difference equations

    Equation (1) gives us the form of the linear difference equation that describes the

    system. Taking z transform on either side and assuming zero initial conditions, we have

    Where H(z) is a z transform of unit sample response h(n).

    Stability analysis

    Jurys stability test

    Bilinear transformation

    Jurys stability test

    Jurys stability testis used to determine whether the roots of the characteristic

    polynomial lie within a unit circle or not. It consists of two parts.One simple test for

    necessary condition for stability and another test for sufficient condition for stability.

    Let us consider a general characteristic polynomial F (z)

    0,................)( 011

    1 n

    n

    n

    n

    n awhereazazazazF

    Necessary condition for stability

    0)1()1(;0)1( FF n

    If this necessary condition is not met, then the system is unstable. We need not check the

    sufficient condition.

  • 7/25/2019 MCS a.pdf

    36/84

    36

    Sufficient condition for stability

    20

    20

    10

    0

    ..................

    rr

    cc

    bb

    aa

    n

    n

    n

    If the characteristic polynomial satisfies (n-1) conditions, then the system is stable

    Jurys test

    Bilinear transformation

    The bilinear transformation maps the interior of unit circle in the z plane into the left half of

    the r-plane.

    1

    1

    z

    zr Or r

    rz

    1

    1

    Fig.Mapping of unit circle in z-plane into left half of r-plane

    Consider the characteristic equation

    )....(..........0;............ 02

    2

    1

    1 iaaazazaza nzn

    n

    n

    n

    n

    n

    Sub r

    r

    z 1

    1

    in Equation (i)

  • 7/25/2019 MCS a.pdf

    37/84

    37

    )..(..........0)1

    1(............)

    1

    1()

    1

    1()

    1

    1( 0

    2

    2

    1

    1 iiar

    ra

    r

    ra

    r

    ra

    r

    ra nn

    n

    n

    n

    n

    Equation (ii) can be simplified

    0............ 012

    2

    1

    1 brbrbrbrb n

    n

    n

    n

    n

    n

    Problem

    1. Check for stability of the sampled data control system srepresented by

    characteristic equation.

    0225)(2 zzi

    Given

    0225)(2 zzzF

    5

    225

    2)1(2)1(5)1(

    225)(

    2

    2

    01

    2

    2

    F

    zzazazazF

    9

    )225(1

    2)1(2)1(5)1()1()1(22F

    n

    Here n=2

    Since 0)1()1(;0)1( FF n , the necessary condition for stability is satisfied.

    Check for sufficient condition

    It consisting of (2n-3) rows

    n=2 (2n-3) = (2*2-3)

    = 1

    So, it consists of only one row

    Row z0 z1 z2

    1 a0 a1 a2

    5,2,2 210 aaa

    10 aa

    The necessary condition to be satisfied

  • 7/25/2019 MCS a.pdf

    38/84

    38

    The necessary & sufficient conditions for stability are satisfied. Hence the system is stable

    (ii) 005.025.02.0)(23 zzzzF

    )(zF 012

    2

    3

    3 azazaza

    005.025.02.0 23 zzz

    Method 1

    Check for necessary condition

    005.025.02.0)( 23 zzzzF

    6.005.0)1(25.0)1(2.01)1(23

    F

    9.0]05.0)1(25.0)1(2.0)1([)1()1()1( 233Fn

    Here n=3

    Since )1(F >0 & )1()1( Fn >0

    The necessary condition for stability is satisfied.

    Check for sufficient condition

    It consisting of (2n-3) row

    n =3, (2n-3) = (2*6-3) =3

    So, the table consists of three rows

    Row z0 z1 z2 z3

    1 a0 a1 a2 a3

    2 a3 a2 a1 a0

    3 b0 b1 b2

    1

    2.0

    25.0

    05.0

    3

    2

    1

    0

    a

    a

    a

    a

  • 7/25/2019 MCS a.pdf

    39/84

    39

    24.0

    )25.0(*)2.0(*05.02.01

    25.005.0

    1875.0

    )2.0*25.0(05.025.01

    2.005.0

    9975.0

    105.005.01

    105.0

    23

    10

    3

    13

    20

    1

    2

    03

    30

    0

    aa

    aab

    aa

    aab

    aa

    aab

    Row z0 z1 z2 z3

    1 0.05 -0.25 -0.2 1

    2 1 -0.2 -0.25 1

    3 -0.9975 0.1875 0.24

    The necessary condition to be satisfied

    25.09975.0,105.0

    , 2030 bbaa

    The necessary and sufficient conditions for stability are satisfied. Hence the system is stable.

    Method 2

    005.025.02.0)( 23 zzzzF

    Putr

    rz

    1

    1

    005.0)1

    1(25.0)1

    1(2.0)1

    1()(

    23

    r

    r

    r

    r

    r

    rrF

    On multiplying throughout by 3)1( r we get

  • 7/25/2019 MCS a.pdf

    40/84

    40

    06.09.26.39.0

    0)2.01.04.03.0()8.08.22.32.1(

    0)2.01.03.02.01.03.0()8.022.18.022.1(

    0)2.01.03.0)(1()8.022.1)(1(

    0)1.005.005.025.025.0)(1()2.02.021)(1(

    0)21)(1(05.0)1)(1(25.0)1)(1(2.0)21)(1(

    0)1(05.0)1)(1(25.0)1()1(2.0)1(

    23

    2323

    232232

    22

    2222

    2222

    3223

    rrr

    rrrrr

    rrrrrrrrrr

    rrrrrr

    rrrrrrrr

    rrrrrrrrrr

    rrrrrr

    The coefficient of the new characteristic equation is positive. Hence the necessary condition

    for stability is satisfied.

    The sufficient condition for stability can be determined by constructing routh array as

    1

    4.........6.0:

    3.........75.2:

    2.........6.06.3:

    1.........9.29.0:

    0

    1

    2

    3

    column

    rowr

    rowr

    rowr

    rowr

    75.26.3

    )6.0*9.0()9.2*6.3(1r

    6.075.2

    )6.3*0()6.0*75.2(0r

    There is no sign change in the elements of first column of routh array. Hence the sufficient

    condition for stability is satisfied.

    The necessary condition and sufficient condition for stability are satisfied. Hence the system

    is stable.

    Pulse transfer function

    It is the ratio of s transform of discrete output signal of the system to the z-transform of

    discrete input signal to the system. That is

    )(

    )()(

    zR

    zCzH (i)

    Proof

    Consider the z-transform of the convolution sum

    k

    k m

    zmrmkhkCZ0 0

    )()()]([ ---------------- (ii)

    On interchanging the order of summation, we get

  • 7/25/2019 MCS a.pdf

    41/84

    41

    k

    km

    zmkhmrzC00

    )(.)()( ------------------ (iii)

    Let mkl Then lkwhenml &0

    0lwhen

    ml

    mkm

    m zlhzmrzC )(.)()(0

    --------------------- (iv)

    l

    mkm

    m zlhzmrzC )(.)()(0

    ------------------------ (v)

    )().()( zHzRzC

    The pulse transfer function

    )(

    )()(

    zR

    zCzH --------------------------- (vi)

    The block diagram for pulse transfer function

    UNIT II

    Z-TRANSFORM AND SAMPLED DATA SYSTEMS

    PART A

    1. What is sampled data control system?

    2. Explain the terms sampling and sampler.

    3. What is meant by quantization?

    4. State (shanons) sampling theorem

    5. What is zero order hold?

    6. What is region of convergence?

    7. Define Z-transform of unit step signal?

    8. Write any two properties of discrete convolution.

    9. What is pulse transfer function?

    10. What are the methods available for the stability analysis of sampled data control

    systems?

    11. What is bilinear transformation?

  • 7/25/2019 MCS a.pdf

    42/84

    42

    PART B

    1. (i)solve the following difference equation

    2 y(k)2 y(k-1) + y (k-2) = r(k)

    y (k) = 0 for k

  • 7/25/2019 MCS a.pdf

    43/84

    43

    UNIT III

    STATE SPACE ANALYSIS OF DISCRETE TIME SYSTEMS

    State variables

    Concepts of State and State Variables

    State

    The state of a dynamic system is the smallest set of variables (called state variables) such

    that the knowledge of these variables at t=t0, together with the knowledge of the inputs for

    0tt ,completely determine the behaviour of the system for any time 0tt .

    The concept of state is not limited to physical systems. It is applicable to biological systems.

    economic systems, social systems, and others.

    State variables

    The state variables of a dynamic system are the smallest set of variables that determine the

    state of the dynamic system. i.e. the state variables are the minimal set of variables such that the

    knowledge of these variables at any initial time t = to, together with the knowledge of the

    inputs for 0tt is sufficient to completely determine the behaviour of the system for any

    time 0tt . If atleast n variables nxxx ,......2,1 are needed to completely describe the

    behaviour of a dynamic system than those n variables are a set of state variables.

    The state variables need not be physically measurable or observable quantities. Variables that

    do not represent physical quantities and those that are neither measurable nor observable can

    also be chosen as state variables. Such freedom in choosing state variables is an added

    advantage of the state-space methods.

    Canonical forms

    They are four main canonical forms to be studied:

    1. Controller canonical form

    2. Observer canonical form

    3. Controllability canonical form

    4. Observability canonical form

    State variables Canonical forms Digitalization Solution of state equations

    Controllability and Observability Effect of sampling time on controllability Pole

    placement by state feedback Linear observer design First order and second order

    problems

  • 7/25/2019 MCS a.pdf

    44/84

    44

    Controller canonical form

    Consider the transfer function of the following for illustration:

    2

    1 2 3

    3 21 2 3

    ( )

    ( )

    b s b s by s

    u s s a s a s a

    The transfer function is firstly decomposed into two subsystems:

    2

    1 2 33 2

    1 2 3

    ( ) ( ) ( ) 1

    ( ) ( ) ( )

    y s y s z sb s b s b

    u s z s u s s a s a s a

    In other words,

    3 2

    1 2 3

    ( ) 1;

    ( )

    z s

    u s s a s a s a

    2

    1 2 3( )andz(s) 1

    b s b s by s

    It is easy to have the state-space equation of

    3 2

    0 1 0 0

    0 0 1 0 ;

    1

    Z Z u

    a a a

    3 1 2 2 1 3 3 2 1

    + b z + b z = b b b Zy b z

    Thus for a general transfer function of

    1

    1 0

    1

    1 0

    ( ) ( )

    m m

    m m

    n n

    n n

    b s b s by s u s

    a s a s a

    The state-space representation can be given as

    0 1

    0 1 0 0

    0 0 1 0

    - - n

    n n n

    A

    a aa

    a a a

    ;

    0

    0

    1

    b

    C= 0 1

    0 0 0

    0 0mb bb

    Ca a a

    For convenience, we shall let 0 1 and 1a m n .

  • 7/25/2019 MCS a.pdf

    45/84

    45

    In other words, for system of the following:

    1 1

    1 2

    1

    1

    ( ) ( )n n

    n

    n n

    n

    b s b s by s u s

    s a s a

    We have

    1 2 1

    0 1 0 0

    0 0 1 0

    - -n n

    A

    a a a

    ;

    0

    0

    1

    b

    1 1n nC b b b

    Observer canonical form

    Now, we set n=3 for illustration. Bu assuming all initial values are zero, can be written as

    3 2 2

    1 2 3 1 2 3s y a s y a sy a y b s y b sy b y

    1

    1 2 1 1 1 1 2 1

    2 3 2 2 2 1 3 2

    3 3 3 3 1 3

    y x

    x x a y b u a x x b u

    x x a y b u a x x b u

    x a y b u a x b u

    In other words,

    1 1

    2 2

    3

    1 0 b

    0 1 ; b b ; 0 0 1

    0 0 b

    a

    A a C

    a

    In general,

    1

    2

    1 0 0 0

    0 1 0; b= ; C= 1 0 0

    1 0

    0 0 0 1n

    a

    aA

    a

    Controllability canonical form

    Again, use n=3 as illustration, the controllability form is given as:

  • 7/25/2019 MCS a.pdf

    46/84

    46

    1

    3 2 1

    2 3 2 1 1

    1

    0 0 1 1

    1 0 ; b= 0 ; C= b b b 1 0

    0 1 0 1 0 0

    a a a

    A a a

    a

    In general,

    1

    2

    1

    0 0

    1 0

    0 1 0 -

    0 0 1 -

    n

    n

    n

    a

    a

    A a

    a

    ; b=

    1

    1 2 1

    2 3 1

    1

    1

    1

    1 0

    1 0 0 0 0

    1 0 0 0 0

    n n

    n n

    n n

    a a a

    a a a

    b b b

    a

    1 0 0C

    The Observability canonical form

    2 1

    0 1 0 0 0

    0 0 1 0 0

    0 0 0 0 1

    - n n

    A

    a a a

    ;

    b

    1

    1

    1

    2 1 1

    1 2 1

    1 0 0 0 0 0

    1 0 0 0 0

    1 0

    1

    n n

    n

    n n

    ba

    bb

    a a ab

    a a a

    C= 1 0 0 0

    Controllability and Observability

    The dynamics of a linear time (shift)) invariant discrete-time system may be expressed in

    terms state (plant) equation and output (observation or measurement) equation as follows

    Where x(k) an n dimensional slate rector at time t =kT. an r-dimensional control (input)vector y(k). an m-dimensional output vector ,respectively, are represented as

  • 7/25/2019 MCS a.pdf

    47/84

    47

    The parameters (elements) of A, an nn (plant parameter) matrix. B an rn control

    (input) matrix, and C an rm output parameter, D an rm parametric matrix are constants

    for the LTI system. Similar to above equation state variable representation of SISO (single

    output and single output) discrete-rime system (with direct coupling of output with input) can

    be written as

    Where the input u, output y and d. are scalars, and b and c are n-dimensional vectors.

    The concepts of controllability and observability for discrete time system are similar to the

    continuous-time system.

    A discrete time system is said to be controllable if there exists a finite integer n and input

    mu(k); ]1,0[ nk that will transfer any state )0(0 bxx to the state nkatx n n.

    Controllability

    Consider the state Equation can be obtained as

    Equation can be written as

    State x can be transferred to some arbitrary state x" in at most n steps to be if p(U) = rank

    of nBABAABB n ].........[ 12 .

    Thus, a system is controllable if the rank composite (n nr) matrix ].........[ 12 BABAABB n

    is n.

    Observability

    Consider the output Equation can be obtained as

  • 7/25/2019 MCS a.pdf

    48/84

    48

    Thus, we can write

    If rank of

    !hen initial state x(0) can be determined from the most n measurements of the output andinput.

    We can, therefore. State that "A discrete time system is observable if the rank of the

    composite nnm matrix.

    Effect of sampling time on controllability

    We have a continuous-time plant which is to be controlled. The control action may be

    either continuous or discrete and must make the plant behave in a desired manner. If discrete

    control action is thought of, then the problem of selection of sampling interval arises. The

    selection of best sampling interval for a digital control system is a compromise among many

    factors. The basic motivation to lower the sampling rate 1/T is the cost. A decrease in

    sampling rate means more time is available for control calculations, hence slower computers

    are possible for a given control function or more control capacity is available for a given

    computer. That economically, the best choice is the slowest possible sampling rate that meets

    all the performance specifications.On the other hand, if the sampling rate is too low, the

    sampler discards part of the information present in a continuous .tirne signal. The

    minimum sampling rate or frequency has a definite relationship with the highest

    significant signal frequency (i.e., signal bandwidth). This relationship is given by the

    Sampling Theorem according to which the information contained in a signal is fully

    preserved in its sampled version so long as the sampling frequency is at least twice the

    highest significant frequency contained in the signal. This sets an absolute lower bound to

    the sample rate selection.

  • 7/25/2019 MCS a.pdf

    49/84

    49

    We are usually satisfied with the trial and error method of selection of sampling interval.

    We compare the response of the continuous-time plant with models discretized for

    various sampling rates. Then the model with the slowest sampling rate which gives a

    response within tolerable limits is selected for future work. However, the method is not

    rigorous in approach. Also a wide variety of inputs must be given to each prospective

    model to ensure that it is a tree representative of the plants.

    Pole placement by state feedback

    Consider a linear dynamic system in the state space form

    In some cases one is able to achieve the goal by using the full state feedback, which

    represents a linear combination of the state variables, that is

    So that the closed loop system given by

    has the desired specifications.

    If the pair (A,b) is controllable, the original system can be transformed into phase variable

    canonical form,i.e it exists a nonsingular transformation of the characteristic polynomial of A

    that is

    Such that

    Where ai are coefficients of the characteristic polynomial of A, that is

    For single input single output systems the state feedback is given by

  • 7/25/2019 MCS a.pdf

    50/84

    50

    After

    Linear observer design

    In a linear time invariant observer for reconstruction of the crystal radius from the weighing

    signal is derived. As a starting point, a linear approximation of the system behaviour can beused. For this purpose the nonlinear equations required for observer design need to be

    linearized around some operating or ( steady state) values, i.e. the equations are expanded in a

    Taylor series which is truncated at the second order

    Can be approximated by

    Around some fixed values00

    , ee va . With new coordinates )tan(00

    cccc vrr

    In the same way one can continue with the remaining equations needed for describing the

    process dynamics. For example, The linear model he derived is

    Where x is the state vector, Furthermore. One has the 3 3 system matrix A. the 3 2 control

    matrix B and the 1 3 output matrix C. One has to keep in mind that the values of the state

    spare vector .r, rho input sector it and the output y describe the deviation of the corresponding

    quantities from their operating willies.

  • 7/25/2019 MCS a.pdf

    51/84

    51

    UNIT III

    STATE SPACE ANALYSIS OF DISCRETE TIME SYSTEMS

    PART A

    1.

    What is state and state variable?

    2. What is a state vector?

    3. What is state space?

    4. What is input and output space?

    5.

    What are the advantages of state space modeling using physical variable?

    6. What are phase variables?

    7. What is the advantage and the disadvantage in canonical form of state model?

    8.

    Write the solution of discrete time state equation?

    9. Write the expression to determine the solution of discrete time state equation using z-

    transform

    10.

    Write the state model of the discrete time system?

    PART

    1. A linear second order single input continuous time system is described by the

    following set of differential equations.

    )()()(2)(

    )(4)(2)(

    21

    .

    2

    21

    .

    1

    tutXtXtX

    tXtXtX

    Comment on the controllability and stability of the system.

    2. The state space representation of a second order system is

    utxxx

    utxx

    )(2

    )(

    21

    .

    2

    1

    .

    1

    State whether the system is controllable.

    3. A system is described by

    XY

    UXX

    01

    1

    0

    11

    11.

    Check the controllability and observability of the system

    4. A control system has a transfer function given by G(s) =2

    )2)(1(

    3

    ss

    s

  • 7/25/2019 MCS a.pdf

    52/84

    52

    Unit IV

    NONLINEAR SYSTEMS

    Introduction to Nonlinear Systems

    It has been mentioned earlier that a control system is said lo be linear ifit obeys law of superposition. Most of the control systems are nonlinear in

    nature and are treated to be linear, under certain approximation, from case of

    analysis point of view. Let us discuss now the properties of nonlinear systems.

    In practice nonlinearities may exist in the systems inherently or may be

    purposely introduced in the systems, to improve the performance. Hence

    knowledge of properties of nonlinear systems and various nonlinearities

    is important.

    Properties of Nonlinear Systems

    The various characteristics of nonlinear systems are,

    The most important characteristics of a nonlinear system is that it does not

    obey the law of superposition. Hence its behavior with respect to standard test

    inputs cannot be used as base to analyses its behavior with respect to other

    inputs. Its response is different for different amplitudes of input signals. Hence

    while doing the analysis of nonlinear system, along with the mathematical model

    of the system, it is necessary to have information about amplitudes of the

    probable inputs, initial conditions etc. This makes the analysis of the nonlinear

    system difficult.

    Linear system gives sinusoidal output for a sinusoidal input, may be

    introducing aphase shift. But nonlinear system produces higher harmonics and

    sometimes the sub harmonics. Hence for sinusoidal input, the output of a

    nonlinear system is generally non sinusoidal. The output consists of frequencies

    which are multiples of the input frequency i.e. harmonics. The sub harmonics

    Types of nonlinearityTypical examplesPhase-plane analysisSingular pointsLimit

    cyclesConstruction of phase trajectoriesDescribing function methodBasic concepts

    Dead ZoneSaturationRelayBacklashLiapunov stability analysisStability in the sense

    of LiapunovDefiniteness of scalar functionsQuadratic formsSecond method of Liapunov

    Liapunov stability analysis of linear time invariant systems and non-linear system

  • 7/25/2019 MCS a.pdf

    53/84

    53

    means the presence of frequencies which are lower than the input

    frequency. The input and output relations are not linear.

    In linear system,the sinusoidal oscillations depend on the input amplitude and

    the initial conditions. But in a nonlinear system, the periodic oscillations may

    exist which are not dependent on the applied input and other system

    parameter variations. In nonlinear system, such periodic oscillations are

    nonsinusoidal having fixed amplitude and frequency. Such oscillations arc

    called limit cycles in case of nonlinear system.

    Another important phenomenon which exists only in case of nonlinear system

    is jump resonance. This can be explained by considering a frequency response.

    The Fig. (a) Shows the frequency response of a linear system which shows

    that output varies continuously as the frequency changes. Similarly though

    frequency us increased or decreased, the output travels along the same curve

    again and again. But in case of a nonlinear system, if frequency is increased,

    the output shows discontinuity i.c. it jumps at a certain frequency. And if

    frequency is decreased, it jumps back but at different frequency. This is shown

    in the fig.

    There is no definite criterion for judging the stability of the nonlinear

    system. The analysis and design techniques of linear systems cannot be

    applied to the nonlinear system.

    Types of nonlinearities

    The nonlinearities can be classified as incidental and intentional.

    The incidental nonlinearities are those which are inherently present in the system.

    Common examples of incidental nonlinearities are saturation, dead zone,

  • 7/25/2019 MCS a.pdf

    54/84

    54

    coulomb friction, stiction, backlash, etc.

    The intentional nonlinearities are those which are deliberately inserted in the

    system to modify system characteristics. The most common examples of this type

    of nonlinearity is a relay.

    In many cases the system presents a nonlinear phenomenon which is fully

    characterised by it static characteristics, i.e., its dynamics can be neglected

    Saturation

    In this type of nonlinearity the output is proportional to input for a limited

    range of input signals. When the input exceeds this range, the output tends to become

    nearly constant as shown in the fig.

    Saturation

    Deadzone

    The deadzone is the region in which the output is zero for a given input. Many

    physical devices do not respond to small signals, i.e., if the input amplitude is less than

    some small value, there will be no output. The region in which the output is zero is called

    deadzone. When the input is increased beyond this deadzone value, the output will be

    linear.

    Dead zone

  • 7/25/2019 MCS a.pdf

    55/84

    55

    Friction

    Friction exists in any system when there is relative motion between contacting surfaces.

    The different types of friction are viscous friction, coulomb friction and stiction.

    Stiction

    The viscous friction is linear in nature and the frictional force is directly proportional to

    relative velocity of the sliding surface.

    Relay

  • 7/25/2019 MCS a.pdf

    56/84

    56

    Phase plane analysis

    Objectives:

    - Use eigenvalues and eigenvectors of the Jacobian matrix to characterize the phase

    plane behavior.

    - Predict the phase-plane behavior close to an equilibrium point, based on the

    - Linearized model at that equilibrium point.

    - Predict qualitatively the phase-plane behavior of the nonlinear system, when there

    are multiple equilibrium points.

    Phase-plane analysis

    Phase plane analysis is a graphical method for studying second-order systems. This

    chapters objective is to gainfamiliarity of the nonlinear systems through the simplegraphical method.

    Concepts of Phase Plane Analysis

    Phase portraits

    The phase plane method is concerned with the graphical study of second-order autonomous

    systems described by

    Where

    x1, x2 : states of the system

    f1, f2: nonlinear functions of the states

    Geometrically, the state space of this system is a plane having x1, x2 as

    coordinates. This plane is called phase plane. The solution of (2.1) with time varies from zero

    to infinity can be represented as a curve in the phase plane. Such a curve is called a phaseplane trajectory. A family of phase plane trajectories is called a phase portrait of a system.

    Example1

    Phase portrait of a mass-spring system as shown in the fig.

    Solution

    The governing equation of the mass-spring system in Fig (a) is the familiar linear second-

    order differential equation

  • 7/25/2019 MCS a.pdf

    57/84

    57

    The governing equation of the mass-spring system in Fig (a) is the familiar linear second-

    order differential equation

    Assume that the mass is initially at rest, at lengthx0. Then the solution of this equation is

    Eliminating time t from the above equations, we obtain the equation of the trajectories

    This represents a circle in the phase plane. Its plot is given in fig (b)

    The nature of the system response corresponding to various initial conditions is directly

    displayed on the phase plane. In the above example, we can easily see that the system

    trajectories neither converge to the origin nor diverge to infinity. They simply circle around

    the origin, indicating the marginal nature of the systems stability. A major class of second-

    order systems can be described by the differential equations of the form

  • 7/25/2019 MCS a.pdf

    58/84

    58

    In state space form, this dynamics can be represented withx1 =x andx2 =x& as follows

    Singular points

    A singular point is an equilibrium point in the phase plane. Since equilibrium point is defined

    as a point where the system states can stay forever,

    Example 2

    A nonlinear second-order system

    The system has two singular points, one at (0,0) and the other at (3,0) . The motion

    patterns of the system trajectories in the vicinity of the two singular points have different

    natures. The trajectories move towards the pointx = 0 while moving away from the pointx =

    3.

    Constructing Phase Portraits

    There are a number of methods for constructing phase plane trajectories for linear or

    nonlinear system, such that so-called analytical method, the method of isoclines, the delta

    method, Lienards method, and Pells method.

    Analytical method

  • 7/25/2019 MCS a.pdf

    59/84

    59

    There are two techniques for generating phase plane portraits analytically. Both

    technique lead to a functional relation between the two phase variablesx1 andx2 in the form

    g(x1,x2 ) = 0 (2.6) where the constant c represents the effects of initial conditions (and,

    possibly, of external input signals). Plotting this relation in the phase plane for different initial

    conditions yields a phase portrait.

    The first technique involves solving (2.1) forx1 andx2 as a function of time t , i.e.,x1(t) =

    g1(t) andx2 (t) =g2 (t) , and then, eliminating time t from these equations. The second

    technique, on the other hand, involves directly eliminating the time variable, by noting that

    and then solving this equation for a functional relation betweenx1 andx2 . Let us use this

    technique to solve the mass spring equation again.

    The first case corresponds to a node.

    Stable or unstable node (Fig.a -b)

    A node can be stable or unstable:

    1,2 < 0 : singularity point is calledstable node.

    1,2 >0 : singularity point is called unstable node.

    There is no oscillation in the trajectories.

    Saddle point (Fig.c)

    The second case (1 < 0 < 2)corresponds to a saddle point. Because of the unstable pole 2

    almost all of the system trajectories diverge to infinity.

  • 7/25/2019 MCS a.pdf

    60/84

    60

    Stable or unstable locus (Fig.d-e)

    The third case corresponds to a focus.

    Re(1,2 ) < 0 : stable focus

    Re(1,2 ) > 0 : unstable focus

    Center point (Fig.f)

    The last case corresponds to a certain point. All trajectories are ellipses and the singularity

    point is the centre of these ellipses.

    Note that the stability characteristics of linear systems are uniquely determined by the

    nature of their singularity points. This, however, is not true for nonlinear systems.

  • 7/25/2019 MCS a.pdf

    61/84

    61

    Limit cycle

    In the phase plane, a limit cycle is defined as an isolated closed curve. The trajectory

    has to be both closed, indicating the periodic nature of the motion, and isolated, indicating the

    limiting nature of the cycle (with nearby trajectories converging or diverging from it).

    Depending on the motion patterns of the trajectories in the vicinity of the limit cycle, we can

    distinguish three kinds of limit cycles.

    Limit cycle can be a drawback in control systems:

    Instability of the equilibrium point

    Wear and failure in mechanical systems

    Loss of accuracy in regulation

    Stable Limit Cycles: all trajectories in the vicinity of the limit cycle converge to it as t

    (Fig.a).

    Unstable Limit Cycles: all trajectories in the vicinity of the limit cycle diverge to it as

    t (Fig.b)

    Semi-Stable Limit Cycles: some of the trajectories in the vicinity of the limit cycle

    converge to it as t (Fig.c)

    Difference between center and limit cycle

    Center trajectories can be found in the linear or linearized systems with the largest real part of

    the eigenvalues of zero value (in marginal point.)

    - depend on the initial conditions

    Limit cycle can occur in nonlinear systems:

    -

    isolated closed orbit(related to Hopf bifurcation)

  • 7/25/2019 MCS a.pdf

    62/84

    62

    Center Limit Cycle

    Describing function method

    . .

    Of all the analytical methods developed over the years for nonlinear systems. the describingfunction method is generally agreed upon as being the most practically useful. It is an

    approximate method, but experience with real systems and computer simulation results, shows

    adequate accuracy in many cases. The method predicts whether limit cycle oscillations will exist

    or not, and gives numerical estimates of oscillation frequency and amplitude when limit cycles

    are predicted. Basically, the method an approximate extension of frequency-response methods

    (including Nyquist stability criterion) to nonlinear systems.

    To discuss the basic concept underlying the describing function analysis. Let us consider the

    block diagram of a nonlinear system shown in Fig. 9.5. Where the blocks GO) and G2(s) represent

    the linear elements. While the blockN represent, the nonlinear element.

    The describing function method provides a "linear approximation" to the nonlinear element

    based on the assumption that the input to the nonlinear element so sinusoid of known,

    constant amplitude. The fundamental harmonic of the element's output is compared with the

    input sinusoid, to determine the steady-state amplitude and phase relation. This relation is the

    describing function for the nonlinear element. The method can, thus, be viewed as "harmonic

    linearization" of a nonlinear element.

    The describing function method is based on the Fourier series. A review of the Fourier series

    will be in order here.

  • 7/25/2019 MCS a.pdf

    63/84

    63

    Fourier series

    We begin with the definition of a periodic signal. A signal y(t) is said to be periodic with the

    period if y(t+T) =y(t) for every value of t. The smallest positive value of T for whichy(t + )

    = y(t) is called fundamental period ofy(t). We denote the fundamental period as T0.Obviously,

    2T0 is also a period of y(t), and so is any integer multiple of T0.A periodic signaly(t) may be

    represented by the series

    The term for n = 1 is called fundamental or first- harmonic, and always has the same

    frequency as the repetition rate of the original periodic waveform; whereas n = 2, 3....,

    give second, third. and so forth harmonic frequencies as integer multiples of the

    fundamental frequency.

    Certain simplifications are possible when y(t) has a symmetry clone type or another.

    The describing Function approach to the analysis of steady-state oscillations in nonlinear

    systems is an approximate tool to estimate the limit cycle parameters.

    It is based on the following assumptions

  • 7/25/2019 MCS a.pdf

    64/84

    64

    There is only one single nonlinear component

    The nonlinear component is not dynamical and time invariant

    The linear component has low-pass filter properties

    The nonlinear characteristic is symmetric with respect to the origin

    There is only one single nonlinear component

    The system can be represented by a lumped parameters system with two main blocks:

    The linear part

    The nonlinear part

    The nonlinear component is not dynamical and time invariant

    The system is autonomous.

    All the system dynamics is concentrated in the linear part

    So that classical analysis tools such as Nyquist and Bode plots can be applied.

    The linear component has low-pass filter properties. This is the main assumption that allows

    for neglecting the higher frequency harmonics that can appear when a nonlinear system is

    driven by a harmonic signal

    The more the low-pass filter assumption is verified the more the estimation error affecting the

    limit cycle parameters is small.

  • 7/25/2019 MCS a.pdf

    65/84

    65

    The nonlinear characteristic is symmetric with respect to the origin. This guarantees that the

    static term in the Fourier expansion of the output of the nonlinearity, subjected to an

    harmonic signal, can be neglected

    Such an assumption is usually taken for the sake of simplicity, and it can be relaxed.

    Ideal relay

    The negative reciprocal of the DF is the negative real axis in backward direction. A limit

    cycle can exist if the relative degree of G(j) is greater than Two

    The oscillation frequency is the critical frequency c of the linear system and the

    Oscillation magnitude is proportional to the relay gain M.

    Liapunovs Stability Analysis

    The state equation for a general time invariant system has the form x = f (x,

    u). If the inputu is constant then the equation will have form x = F(x).

    For this system, the points, at which derivatives of all state variables are

    zero, are the singular points.

    These singularpoints are nothing but equilibrium points where the system

    stays if it is undisturbed when the system is placed at these points.

  • 7/25/2019 MCS a.pdf

    66/84

    66

    The stability of such a system is defined in two different ways. If the

    input to the system is zero with arbitrary initial conditions, the resulting

    trajectory in phase-plane, discussed in earlier chapter, tends towards the

    equilibrium state.

    If tile input to the system is provided then the stability is defined as for

    bounded input, the system output is also bounded.

    For linear systems with non-zero eigen values, there is only one

    equilibrium state and the behaviour of such systems about this

    equilibrium state totally determines the qualitative behaviour in the

    entire state space.

    In case of nonlinear systems, the behaviour for small deviations about the

    equilibrium point is different from that for large deviations.

    Hence local stability for such systems does not indicate the overall

    stability in the state space. Also the non-linear systems having multiple

    equilibrium states, the trajectories move from one equilibrium point and

    tend to other with time.

    Thus stability in case of non-linear system is always referred toequilibrium state instead of global term stability which is the total stability

    of the system.

    In case of linear control systems, many of the stability criteria such as

    Routh's stability test. Nyquist stability criterions etc. are available. But

    these cannot be applied (or non-linear systems.

    The second method of Liapunov which is also called direct method of

    Liapunov is the most common method for obtaining the stability of non-

    linear systems.

    This method is equally applicable to time varying systems, stability

    analysis of linear, time in variant systems and for solving quadratic

    optimal control problem.

    Stability in the Sense of Liapunov

    Consider a system defined by the state equation ),(.

    txfx .Let us assume

    that this system has a unique solution starting at the given initial condition. Let us

  • 7/25/2019 MCS a.pdf

    67/84

    67

    consider this solution as ):( 0,0txtF where 0xx at 0tt and t is the observed time.

    0000 ),:( xtxtF

    If we consider a state ex for system with equation ),(.

    txfx in such a way that

    0),( txf e for all t then this ex is called equilibrium state. For linear, time

    invariant systems having A non-singular, there is only one equilibrium state while

    there are one ormore equilibrium states if A is singular.

    In case of non-linear systems as we have seen previously there are more than

    one equilibrium states. The Isolated equilibrium states that is isolated from each

    other can be shifted to origin i.e. f(0, t) = 0 by properly shifting the coordinates.

    These equilibrium states can be obtained from the solution of equation f(x.. t) = 0.

    Now we will consider the stability analysis of equilibrium states at the origin. We

    will consider a spherical region of radius R about an equilibrium state ex ,. Such that

    Any equilibrium state ex of the system ),(.

    txfx is said to be stable in the

    sense of Liapunov if corresponding to each S( ) there is S( )such that trajectories

    staring in S( ) do not leave S( ) as time t increases indefinitely. The real

    number depends on and in general also depends on t0. If does not depend on

    t0, the equilibrium state is said to be uniformly stable.

    The region S( ) must be selected first and for each S( ) , there must be a

    region S( ) in such a way that the trajectories staring within S( ) do not leave S( )

    as time t progresses.

    There are many types of stability definitions such as asymptotic stability,

    asymptotic stability in large. We will also see the definition of instability along

    with definitions of these types of stability.

  • 7/25/2019 MCS a.pdf

    68/84

    68

    Definiteness of scalar function

    Positive Definiteness

    A scalar function F(x) is said to be positive definite in a particularregion which includes the origin of state space if F(x) > 0 for all non-zero states

    x in that region and F(0) = 0.

    Negative Definiteness

    A scalar function F(x) is said to be negative definite if - F(x) is positive

    definite.

    Positive Semidefiniteness

    A scalar function F(x) is said to be positive semi definite if it is positive at all states in

    the particular region except at the origin and at certain other states where it is zero.

    Negative SemidefiniteA scalar function F(x) is said to be negative semidefinite if - F(x) is positive

    semidefinite.

    Indefiniteness

    A scalar function F(x) is said to be indefinite in the particular region if it

    assumes both positive and negative values irrespective how small the region is

    Quadratic Form

    A class of scalar functions which plays important role in Ihe stability analysis based on

    Liapunov's second method is the quadratic form

  • 7/25/2019 MCS a.pdf

    69/84

    69

    P is real symmetric matrix and x is a real vector.

    Liapunov's Second Method

    A system which is vibrating is stable if its total energy is continuously

    decreasing. This indicates that the lime derivative of the total energy must

    be negative.

    The energy is decreased till an equilibrium state is reached. The total

    energy is a positive definite function

    This fact obtained from classical mechanics theory is generalized in Liapunov's

    second method. If the system has an asymptotically stable equilibrium stale then Ihe

    stored energy decays with increase in time till it attains minimum value at the

    equilibrium state.

    But there is no simple way for defining an energy function. For purely mathematical

    system. This difficulty was overcome as Liapunov introduced Liapunov function

    method which is fictitious energy function.

    Liapunov functions depend on x1, x2, .. xnand t. It is given as F(x1, x2, ... xn, t) or as

    F(x, t). In Liapunov's second method, the sign behaviour of F(x, t) and its time

    derivative F(x, t) = dF(x, t)/dt gives as information about stability, asymptotic

    stability or instability of an equilibrium state without requiring to solve the equations

    directly to get the solution

    Liapunov's Stability Theorem

    Consider a scalar function V(x), where x is n vector and is positive

    definite, then the states x that satisfy V(x) = C, where C is a positive constant, lie

    on a closed hyper surface in n dimensional slate space at least in the

    neighborhood of origin. This is shown in the Fig.

    If V(x) is a positive definite function obtained for a given system such that

    its time derivative taken along the trajectory is always negative then V(x) becomes

    smaller and smaller in terms of C and finally reduced to zero as x reduces to

    zero. This indicates asymptotic stability of the origin. Liapunov's main stability

    theorem is based on this and gives a sufficient condition for asymptotic stability.

  • 7/25/2019 MCS a.pdf

    70/84

    70

    Liapunov's stability theorem is as given below. Consider a system

    described by equation x f(x, t) where f(o, t) = 0 for all L If thereexists a

    scalar function V(x,t) having continuous first partial derivatives and

    satisfying the conditions such as V(x, t) ispositive definite and V(x, t) is

    negative definite then the equilibrium state at the origin is uniformly

    asymptotically stable.

    Consider the system described by x = f(x, t) where f (0, t) = 0 for all If

    there exists a scalar function V(x, t) having continuous first partial derivatives

    and V(x, t) ispositive definite, V (x, t) is negative semidefinite V (0 (t: xg, tg),

    t) docs not vanish identically in t a t for any t0 and any Xg * 0 where 0 (t: Xg,

    tg) denotes or indicates the solution starting from Xg at tg then the

    equilibrium state at origin of the system is uniformly asymptotically stable in

    the large.

    The equilibrium state at origin is unstable when there exists a scalar

    function U(x,t) having continuous, first partial derivatives and satisfying the

    conditions U (x, t) is positive definite in some region about the origin and U (x, t)

    is positive definite in the same region.

    Stability of Linear and Nonlinear Systems

    If the equilibrium state in case of linear, time invariant system is

    asymptotically stable locally then it is asymptotically stable in the large. But

    in case of a nonlinear system, the equilibrium state has to be

    asymptotically stable in the large for the state to be locally asymptotically

    stable. Hence the asymptotic stability of the equilibrium state of linear, time

  • 7/25/2019 MCS a.pdf

    71/84

    71

    invariant systems and those of nonlinear systems is different.If it is required to

    ent the asymptotic stability of any equilibrium state for a nonlinear system

    then the stability analysis of linearized models of non-linear systems is

    totally insufficient. The nonlinear systems are to be tested without making

    them linearized.

    The Direct Method of Liapunov and the Linear System

    For linear systems, Liapunov's direct method proves to be a simple

    method for stability analysts. Use of Liapunov's method for linear systems

    is helpful in extending the thinking towards nonlinear systems.

    Consider linear system described by state equaion

    The linear system described by above equation is asymptotically stable in the

    large at the origin if and only if for any symmetric, posiive definite matrix

    Q, there exists a symmetic posiive definite matix P which is th