On Identifiability of Nonlinear Ode Models And

38
ON IDENTIFIABILITY OF NONLINEAR ODE MODELS AND APPLICATIONS IN VIRAL DYNAMICS HONGYU MIAO * , XIAOHUA XIA , ALAN S. PERELSON , AND HULIN WU § Abstract. Ordinary differential equations (ODE) are a powerful tool for modeling dynamic processes with wide applications in a variety of scientific fields. Over the last 2 decades, ODEs have also emerged as a prevailing tool in various biomedical research fields, especially in infectious disease modeling. In practice, it is important and necessary to determine unknown parameters in ODE models based on experimental data. Identifiability analysis is the first step in determing unknown parameters in ODE models and such analysis techniques for nonlinear ODE models are still under development. In this article, we review identifiability analysis methodologies for nonlinear ODE models developed in the past one to two decades, including structural identifiability analysis, practical identifiability analysis and sensitivity-based identifiability analysis. Some advanced topics and ongoing research are also briefly reviewed. Finally, some examples from modeling viral dynamics of HIV, influenza and hepatitis viruses are given to illustrate how to apply these identifiability analysis methods in practice. Key words. ODE modeling, structural identifiability, practical identifiability, sensitivity-based identifiability, viral dynamics AMS subject classifications. 34A30, 34A34, 34C20 1. Introduction. Ordinary differential equation (ODE) models have been wide- ly used to model physical phenomena, engineering systems, economic behavior, and biomedical processes. In particular, ODE models have recently played a prominent role in describing both the within host dynamics and epidemics of infectious diseases and other complex biomedical processes (e.g., [2, 15, 59, 74, 75, 77]). Great attention has been paid to the so-called forward problem or simulation problem, i.e., predicting and simulating the results of measurements or output variables for a given system with given parameters. However, less effort has been devoted to the inverse problem, i.e., using the measurements of some state or output variables to estimate the pa- rameters that characterize the system, especially for nonlinear ODE models without closed-form solutions. In reality, before rigorous parameter estimation methods can be applied to an ODE model to formally estimate the model parameters based on experimental data, a serious barrier to overcome is how to verify whether the model parameters are identifiable based on the measurements of output variables or their functions when the ODE model does not have a closed-form solution. Further questions include, if not all model parameters are identifiable, are a subset of parameters identifiable? How many measurements, at which time points, are necessary to identify the identifiable parameters? To answer these questions, identifiability analysis should be done before tackling the inverse problem. * Department of Biostatistics and Computational Biology, University of Rochester School of Medicine and Dentistry, 601 Elmwood Avenue, Box 630, Rochester, New York 14642, USA (hongyu [email protected]). Department of Electrical, Electronic and Computer Engineering, University of Pretoria, Lyn- nwood Road, Pretoria 0002, South Africa ([email protected]). Theoretical Biology and Biophysics Group, MS-K710, Los Alamos National Laboratory, Los Alamos, NM 87545, USA ([email protected]). § Department of Biostatistics and Computational Biology, University of Rochester School of Medicine and Dentistry, 601 Elmwood Avenue, Box 630, Rochester, New York 14642, USA ([email protected]). 1

description

Abstract. Ordinary differential equations (ODE) are a powerful tool for modeling dynamicprocesses with wide applications in a variety of scientific fields. Over the last 2 decades, ODEshave also emerged as a prevailing tool in various biomedical research fields, especially in infectiousdisease modeling. In practice, it is important and necessary to determine unknown parametersin ODE models based on experimental data. Identifiability analysis is the first step in determingunknown parameters in ODE models and such analysis techniques for nonlinear ODE models arestill under development. In this article, we review identifiability analysis methodologies for nonlinearODE models developed in the past one to two decades, including structural identifiability analysis,practical identifiability analysis and sensitivity-based identifiability analysis. Some advanced topicsand ongoing research are also briefly reviewed. Finally, some examples from modeling viral dynamicsof HIV, influenza and hepatitis viruses are given to illustrate how to apply these identifiability analysismethods in practice.

Transcript of On Identifiability of Nonlinear Ode Models And

Page 1: On Identifiability of Nonlinear Ode Models And

ON IDENTIFIABILITY OF NONLINEAR ODE MODELS ANDAPPLICATIONS IN VIRAL DYNAMICS

HONGYU MIAO∗, XIAOHUA XIA† , ALAN S. PERELSON‡ , AND HULIN WU§

Abstract. Ordinary differential equations (ODE) are a powerful tool for modeling dynamicprocesses with wide applications in a variety of scientific fields. Over the last 2 decades, ODEshave also emerged as a prevailing tool in various biomedical research fields, especially in infectiousdisease modeling. In practice, it is important and necessary to determine unknown parametersin ODE models based on experimental data. Identifiability analysis is the first step in determingunknown parameters in ODE models and such analysis techniques for nonlinear ODE models arestill under development. In this article, we review identifiability analysis methodologies for nonlinearODE models developed in the past one to two decades, including structural identifiability analysis,practical identifiability analysis and sensitivity-based identifiability analysis. Some advanced topicsand ongoing research are also briefly reviewed. Finally, some examples from modeling viral dynamicsof HIV, influenza and hepatitis viruses are given to illustrate how to apply these identifiability analysismethods in practice.

Key words. ODE modeling, structural identifiability, practical identifiability, sensitivity-basedidentifiability, viral dynamics

AMS subject classifications. 34A30, 34A34, 34C20

1. Introduction. Ordinary differential equation (ODE) models have been wide-ly used to model physical phenomena, engineering systems, economic behavior, andbiomedical processes. In particular, ODE models have recently played a prominentrole in describing both the within host dynamics and epidemics of infectious diseasesand other complex biomedical processes (e.g., [2, 15, 59, 74, 75, 77]). Great attentionhas been paid to the so-called forward problem or simulation problem, i.e., predictingand simulating the results of measurements or output variables for a given systemwith given parameters. However, less effort has been devoted to the inverse problem,i.e., using the measurements of some state or output variables to estimate the pa-rameters that characterize the system, especially for nonlinear ODE models withoutclosed-form solutions.

In reality, before rigorous parameter estimation methods can be applied to anODE model to formally estimate the model parameters based on experimental data,a serious barrier to overcome is how to verify whether the model parameters areidentifiable based on the measurements of output variables or their functions whenthe ODE model does not have a closed-form solution. Further questions include, if notall model parameters are identifiable, are a subset of parameters identifiable? Howmany measurements, at which time points, are necessary to identify the identifiableparameters? To answer these questions, identifiability analysis should be done beforetackling the inverse problem.

∗Department of Biostatistics and Computational Biology, University of Rochester School ofMedicine and Dentistry, 601 Elmwood Avenue, Box 630, Rochester, New York 14642, USA(hongyu [email protected]).

†Department of Electrical, Electronic and Computer Engineering, University of Pretoria, Lyn-nwood Road, Pretoria 0002, South Africa ([email protected]).

‡Theoretical Biology and Biophysics Group, MS-K710, Los Alamos National Laboratory, LosAlamos, NM 87545, USA ([email protected]).

§Department of Biostatistics and Computational Biology, University of Rochester School ofMedicine and Dentistry, 601 Elmwood Avenue, Box 630, Rochester, New York 14642, USA([email protected]).

1

el
Highlight
Page 2: On Identifiability of Nonlinear Ode Models And

2 H. Miao, X. Xia, A.S. Perelson and H. Wu

The literature on ODE identifiability analysis is found in journals from a variety ofscientific fields such as mathematics, biomedical modeling, engineering, and statistics;in addition, various techniques and methodologies from these disciplines are employedin ODE identifiability studies. Therefore, it is useful to have a comprehensive reviewon these methods and approaches, and their further applications, e.g., in experimentaldesign [32, 70, 71, 96]. In this paper, we review identifiability methods with a focuson nonlinear ODE models for which close-form solutions are not available. In Section2, we review various definitions related to identifiability analysis. We review varioustechniques for structural identifiability analysis in Section 3. In addition to theoreti-cal (structural) identifiability, it is also important to evaluate practical identifiabilitywhen experimental data are contaminated with measurement noise. This will be re-viewed in Section 4. Sensitivity analysis is widely used in mathematical modeling toevaluate how sensitive output variables are to parameter values and input variables.Some sensitivity analysis techniques can also be used to evaluate parameter identifi-ability in ODE models, as will be reviewed in Section 5. We illustrate identifiabilitytechniques using examples from infectious disease modeling in Section 6. We concludethis paper with some discussions and summary in Section 7.

2. Definitions. A general dynamic system can be expressed as follows:

x(t) = f(t,x(t),u(t),θ), (2.1)

y(t) = h(x(t),u(t),θ), (2.2)

where x(t) ∈ Rm is a vector of state variables (or dependent variables), y(t) ∈ Rd

the measurement or output vector, u(t) ∈ Rp the known system input vector, andθ ∈ Rq the parameter vector. The system given by Eq. (2.1) is an ordinary differentialequation model (ODE model). For the inverse problems, θ is unknown and has to beestimated based on experimental data. There are three common scenarios for θ:

(i) constant parameters only;(ii) time-varying parameters only;(iii) a mixture of both constant and time-varying parameters.

Let θ = (θc,θt), where θc denotes the constant unknown parameters and θt denotesthe time-varying unknown parameters. Now Eq. (2.1) and (2.2) can be re-written inthe form:

x(t) = f(t,x(t),u(t),θc,θt), (2.3)

y(t) = h(x(t),u(t),θc,θt). (2.4)

Before we introduce definitions of identifiability, we review three important and usefulconcepts developed in control theory: reachable, controllable and observable [60, 111].

Definition 2.1. Reachable: For a certain initial state x0 of interest, a statex1 is said to be reachable if there exists a trajectory x(t) starting from x0 which canachieve x1 in a finite time given an admissible system input u(t).

Note that in the definition above, u(t) is called an admissible input (or admissiblecontrol) if it satisfies all system constraints at any time of interest and a solution ofthe dynamic system exists. The existence of such an admissible u(t) leads to thedefinition of controllability.

Definition 2.2. Controllable: If there exists an admissible u(t) which can trans-fer an initial state of interest to a target state in a finite time, the dynamic system issaid to be controllable.

Page 3: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 3

Controllability is an important property of a dynamic system since it indicateswhether a system will respond to a certain input and behave as expected. One im-portant application of this concept is stabilization of dynamic systems. In biomedicalresearch, the dynamic system could be a virus, such as HIV, infecting a human andthe system input could be antiretroviral therapy; how to control or stabilize the virusvia intentionally-designed intervention strategies is still an interesting and challengingresearch topic [1, 22, 55, 58, 80, 81, 126].

To better understand dynamic system structure and behavior, it is also necessaryto obtain measurements of the system output variables. However, we may not be ableto directly measure the state variables; instead, we may be able to measure outputvariables which are functions of system input and state variables, as specified in Eq.(2.2) or (2.4). If it is possible to determine the system state from system outputmeasurements, the system is observable according to the following definition:

Definition 2.3. Observable: Given an initial state x0 and an admissible controlu(t), if the current system state x(t) can be determined only through the system outputy(t) in a finite time, the system is said to be observable.

In the definition above, it is usually assumed that the system output y(t) canbe measured without error. The three definitions introduced so far describe the rela-tionships among four basic factors of a dynamic system: initial state, input, currentstate and output. There are also other interesting concepts and definitions originat-ing in control theory for connecting these four basic factors of dynamic systems (e.g.,[23, 38, 50, 111]).

The concepts discussed above have been introduced for systems with known pa-rameters. However, these concepts, especially controllability and observability, arealso directly related to system (parameter) identifiability. A system which is control-lable and observable has strong connections among input, state and output variables,and such strong connections may indicate that the system is identifiable. The readeris referred to [4, 19, 20, 98, 99, 101, 110, 123] for further discussions.

Definition 2.4. Identifiable : The dynamic system given by Eqs. (2.1) and(2.2) is identifiable if θ can be uniquely determined from the given system input u(t)and the measurable system output y(t); otherwise, it is said to be unidentifiable.

Limited system inputs may not result in system outputs with sufficient infor-mation for uniquely determining system parameters. Thus, it is also necessary tohave informative input signals in order to identify system parameters. This idea wasformalized by introducing the concept of a persistently exciting input [63, 64, 65].Simply speaking, an input is said to be persistently exciting if enough information onthe output variables can be generated from the input to identify system parametersin the sense that all estimates of system parameters converge to their true values ina finite time [47]. The assumption of persistently exciting inputs is a prerequisite forstructural identifiability analysis [10] as will be discussed in the next section.

Ljung and Glad [64] introduced two important concepts, globally identifiable andlocally identifiable.

Definition 2.5. Globally identifiable: A system structure is said to be globallyidentifiable if for any admissible input u(t) and any two parameter vectors θ1 and θ2

in the parameter space Θ, y(u,θ1) = y(u,θ2) holds if and only if θ1 = θ2.

Definition 2.6. Locally identifiable: A system structure is said to be locallyidentifiable if for any θ within an open neighborhood of some point θ∗ in the parameterspace, y(u,θ1) = y(u,θ2) holds if and only if θ1 = θ2.

Page 4: On Identifiability of Nonlinear Ode Models And

4 H. Miao, X. Xia, A.S. Perelson and H. Wu

Both definitions use the concept of one-to-one mapping between parameters andsystem input/output. With the development of various identifiability analysis tech-niques, more specific definitions of identifiability have been introduced by a numberof authors [7, 10, 21, 56, 64, 114, 119]. For instance, Tunali and Tarn [104] introduceda definition of identifiability when an initial state is given, which was termed locallystrongly identifiable. A similar concept was introduced in [56], called x0-identifiable.

Definition 2.7. Locally strongly identifiable (x0-identifiable) : For an admissibleinput u(t) in the time range of interest [t0, t1] and a given initial state x0 = x(t0)which is independent of θ and not an equilibrium point, if there exists an open setΘ0 within the parameter space Θ such that for any two different parameter vectorsθ1,θ2 ∈ Θ0, the solutions x(t,θ,u) exists on [t0, t0 + ϵ] (t0 < ϵ ≤ t1 − t0) for both θ1

and θ2, and y(t,θ1,x0,u(t)) = y(t,θ2,x0,u(t)) on [t0, t0 + ϵ], the system structureis said to be locally strongly identifiable (or x0-identifiable).

We notice that this definition is specific for differential equation systems. Butit is stringent with respect to the initial state. More generally, Xia and Moog [119]introduced structurally identifiability as follows:

Definition 2.8. Structurally identifiable: Let CNu [t0, t1] denote the function

space expanded by all input functions on [t0, t1] which are differentiable up to theorder N , and let M denote an open set of initial system states. The system structureis said to be structurally identifiable if there exist open and dense subsets M0 ⊂ M ,Θ0 ⊂ Θ, and U0 ⊂ CN

u [t0, t1] such that the system is locally strongly identifiable at θgiven u for any x0 ∈ M0, θ ∈ Θ0, and u ∈ U0.

This definition is also interchangeably called geometrical identifiable [104, 56].Besides these identifiability definitions based on one-to-one mappings between systemparameters and system input-output, Glad [37] and Ljung and Glad [64] introduced adefinition of identifiability based on the algebraic equations consisting of the systeminput and output, which was called algebraically identifiable. This definition is directlyrelated to identifiability analysis techniques [29, 64, 118, 119]:

Definition 2.9. Algebraically identifiable: Based on algebraic equations of sys-tem state, input and output, if a meromorphic function

Φ = Φ(θ,u, u, ...,u(k),y, y, ...,y(k)), Φ ∈ Rq,

can be constructed after a finite number of steps of algebraic calculation or differenti-ation such that Φ = 0 and det∂Φ∂θ = 0 hold in the time range of interest [t0, t1] for any(θ,x0,u) in an open and dense subset of Θ × M × CN

u [t0, t1], where k is a positiveinteger, u, ...,u(k) the derivatives of u, and y, ...,y(k) the derivatives of y, the systemstructure is said to be algebraically identifiable.

Similarly algebraically identifiability with known initial conditions can be definedas follows [119]:

Definition 2.10. Algebraically identifiable with known initial conditions:If ameromorphic function Φ = Φ(θ,x0,u(t

+0 ), u(t

+0 ), ...,u

(k)(t+0 ),y(t+0 ), y(t

+0 ), ...,y

(k)(t+0 )),Φ ∈ Rq, can be constructed from algebraic equations of system state, input and outputafter a finite number of steps of algebraic calculation or differentiation such that Φ = 0and det∂Φ∂θ = 0 hold for all (θ,x0,u(t

+0 ), u(t

+0 ), ...,u

(k)(t+0 ),y(t+0 ), y(t

+0 ), ...,y

(k)(t+0 )),

where k is a positive integer, (θ,x0,u(t+0 ), u(t

+0 ), ...,u

(k)(t+0 )) in an open and densesubset of Θ ×M × U , (u(t+0 ), u(t

+0 ), ...,u

(k)(t+0 )) and (y(t+0 ), y(t+0 ), ...,y

(k)(t+0 )) arethe derivatives of u and y at time t+0 , respectively, the system structure is said to bealgebraically identifiable with known initial conditions.

A number of studies have considered system identifiability given initial conditions[26, 37, 64, 83, 104, 119] and reported that known initial conditions can help to

Page 5: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 5

identify more parameters. Particularly, the work of Wu et al. [117] clarified that,giving initial conditions is equivalent to have more observations on system outputsuch that parameter estimation reliability can be improved, especially for dynamicsystems sensitive to initial conditions.

3. Structural Identifiability Analysis. In this section, we will review struc-tural identifiability methods in details. We will also discuss the advantages and disad-vantages of these methods in practical applications in order to help the practitionerschoose the appropriate approach for specific problems. Furthermore, we will discussthe minimum number of observations obtained via structural identifiability analysisto uniquely determine all identifiable parameters, keeping in mind that this numbercould be much higher for real problems due to the presence of measurement error ormodel uncertainty.

The concept of structural identifiability was first introduced by Bellman andAstrom [10]. As suggested by its name, the corresponding techniques verify sys-tem identifiability by exploring the system structure (that is, the model itself). Earlystructural identifiability analysis techniques were developed from control theories in1970s for linear models, especially compartmental models. For instance, Bellman andAstrom [10] proposed an analysis technique for linear ODE models based on Laplacetransforms. Later, the method of power series expansion was proposed by Pohjanpalo[84] and the similarity transformation method was proposed by Walter and Lecourtier[113] for linear ODE models. These methods have been well summarized in [3] and[51]. In this paper, we focus on identifiability methods for nonlinear ODE modelsinstead of linear models.

Some of the approaches for linear ODE models such as the similarity transforma-tion method have been extended by Vajda and Rabitz [107], Vajda et al. [106] andChappel and Godfrey [16] to nonlinear ODE models. However, the extension onlyworks for a limited number of simple nonlinear problems [7]. For general nonlinearmodels, new techniques are needed. A simple approach for this purpose, called directtest , was proposed by Denis-Vidal and Joly-Blanchard [25] and Walter et al. [112].The basic idea of this approach is to directly use the identifiability definition to verifyparameter identifiability, either analytically [25] or numerically [112]. However, theanalytical direct test is not suitable for high-dimensional problems and the numericaldirect test also has some limitations due to the use of a cut-off value.

Under the framework of differential algebra [91], new methods and algorithmshave also been developed to target identifiability of general nonlinear models [14, 64,78]. The differential algebra approach can utilize the power of symbolic computations,which requires much less human intervention. Since the differential algebra methodwas introduced to investigate the structural identifiability problem [64, 78] in theearly 1990s, it has been successfully applied to nonlinear differential equation models,including models with time-varying parameters [7]. Ljung and Glad [64] summarizedthree conditions under which the system structure is globally identifiable, locallyidentifiable or not identifiable, respectively; however, to verify the three conditions,rigorous mathematical theories need to be further developed.

Xia and Moog [119] proposed another method based on the Implicit Function The-orem. By taking derivatives of observable system outputs with respect to independentvariables (e.g., time), all latent variables (unobservable system state variables) canbe eliminated after algebraic calculations and a finite number of equations consistingof known system inputs, observable system outputs and unknown parameters can beformed. Then a matrix (called the identification matrix) consisting of the partial

el
Highlight
Page 6: On Identifiability of Nonlinear Ode Models And

6 H. Miao, X. Xia, A.S. Perelson and H. Wu

derivatives of these equations with respect to unknown parameters (and usually theirderivatives with respect to independent variables) can be formed. If the identificationmatrix is non-singular, this system is identifiable. This method has the advantages oftheoretical and practical simplicity and has been successfully applied to HIV dynamicmodels of up to 6-dimensions [70, 119]. However, this method requires high orderderivatives, thus the matrix can easily become very complicated and the singularityof the matrix becomes difficult to verify. Wu et al. [117] further extended this methodby considering multiple time points instead of high order derivatives to overcome thedisadvantages of Xia’s method. The methods based on the Implicit Function Theoremcan be applied alone to verify system identifiability and they can also be employed toverify the three conditions in the differential algebra approach. However, for dynamicmodels with time-varying parameters, the singularity of the identification matrix isdifficult to evaluate and no reliable conclusion can be easily drawn. Therefore, inpractice, the differential algebra approach and the Implicit Function Theorem ap-proach may have to be combined to solve a problem. In addition, if initial conditionsare unknown, the correlation between unknown initial conditions and other model pa-rameters can not be verified by structural identifiability analysis unless such unknowninitial conditions explicitly appear on the right-hand side of Eq. (2.1).

Before moving onto the technical details, it is also helpful to mention it here thatstructural identifiability analysis methods are not widely used in practice, yet, due toeither the computational complexity or the lack of mature computer implementations.

3.1. Power Series Expansion and Similarity Transformation. Grewal andGlover [39] studied the identifiability problem for nonlinear ODE models by consid-ering local linearization of nonlinear systems. However, “the linearized system beingidentifiable” is only a necessary condition for “the nonlinear system being identi-fiable” instead of a sufficient condition. Therefore, the local linear approximationcannot answer the question completely. Pohjanpalo [84] proposed another approachcalled power series expansion to better handle nonlinear problems.

For the power series expansion method, the function f in Eq. (2.3) is assumedto be infinitely differentiable with respect to time t, u and x in the time range ofinterest [t0, t1]; and the same assumption is needed for x, y and u with respect totime, and for h with respect to x. Such assumptions are necessary because the powerseries expansion may require derivatives of arbitrary orders. The nonlinear systemconsidered by Pohjanpalo [84] is of the following form:

x = A(t,x,θc)x+ u, (3.1)

y = C(θc)x, (3.2)

which is very restrictive. Consider the derivatives of system output y at time t0ak(t0) = y(k)(t0),

where k denotes the kth derivative of y. Therefore, the system input and output canbe connected by their derivatives with respect to time at t0:

Cx(t0) = a0(t0), (3.3)

C

[k∑

i=1

(k − 1)!

(k − i)!(i− 1)!A(k−i)(t0)x

(i−1)(t0) + u(k−1)(t0)

]= ak(t0), (3.4)

where k = 1, ...,∞. Since the derivatives of y are theoretically observable, they areconsidered as known. Therefore, an infinite number of equations can be obtained from

Page 7: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 7

Eq. (3.3) and (3.4) to solve for θc simultaneously. If the solution is unique, then thesystem structure is (locally) identifiable.

In nature, the power series expansion method is an approach to verify the x0

identifiability (or local strong identifiability). However, this method has a seriousdrawback: high order derivatives are needed for a high dimensional problem and theresulting equations can easily become too complicated to solve. This disadvantagehas prevented this method from becoming popular in practice.

Walter and Lecourtier [113] initially proposed the similarity transformation methodfor linear ODE models. The system concerned here is in the form:

x = Ax+Bu, (3.5)

y = Cx, (3.6)

where A, B and C are matrices of constant coefficients. The basic idea of this methodis to find the similar matrix S = P−1AP of A such that

x =(P−1AP

)x+Bu, (3.7)

y = Cx, (3.8)

where P is a non-singular matrix. It is straight forward to show that if the onlypossible similar transformation of A is P = I, the system is uniquely and globallyidentifiable; if a finite number of P = I can be found, the system is locally identifiable(or nonuniquely identifiable); otherwise, the system is unidentifiable.

Vajda et al. [106] and Vajda and Rabitz [107] extended the work of Walterand Lecourtier [113] and proposed the similarity transformation method to tackle thenonlinear ODE systems by making use of the local state isomorphism theorem [44, 50].The nonlinear system considered in Vajda et al. [106] is of the following form:

x = f (x(t,θ),θ) + u(t)g (x(t,θ),θ) , (3.9)

y = h(x(t,θ),θ). (3.10)

Note that although this system is a single-input system, the conclusion based on thissystem can be generalized to multiple-input systems. It is necessary to introduce thedefinition of structural equivalence before we further introduce the similarity trans-formation method.

Definition 3.1. Structural equivalence: Given two systems of the family inEq. (3.9) and (3.10), if there exist two parameters θ1,θ2 ∈ Θ such that for the sameadmissible input u(t), the solution of the two system exists for θ1 and θ2, respectively,and the corresponding system outputs are the same, the system with parameter θ1 issaid to be equivalent to the system with parameter θ2, denoted by θ1 ∼ θ2.

Under the similarity transformation framework, the identifiability problem be-comes a system equivalence problem: a system structure is identifiable if no equivalentsystems exist for θ1,θ2 ∈ Θ and θ1 = θ2 [106].

Knowledge about Lie algebra is needed to better understand the similarity trans-formation method; however, Lie algebra itself is a very rich topic, which will not beintroduced in detail here. The interested reader is referred to [36]. Based on the workof Hermann and Krener [44], Vajda et al. [106] eventually proposed the similaritytransformation approach to verify the global identifiability, for which a set of partialdifferential equations need to be formed and solved.

In summary, before the similarity transformation method can be applied, it isrequired that the system is both controllable and observable. Further more, a set

Page 8: On Identifiability of Nonlinear Ode Models And

8 H. Miao, X. Xia, A.S. Perelson and H. Wu

of partial differential equations need to be generated and solved [106] to verify thesystem identifiability. These two disadvantages make the similarity transformationmethod not feasible for general nonlinear systems in practice.

3.2. Direct Test. Recall the definition of global (or local) identifiability, the keyis to verify whether the same system output will result in a unique set of parametervalues. That is,

y(u,θ1) = y(u,θ2) ⇒ θ1 = θ2

should be satisfied either globally or locally if the model is identifiable. Based onthis sufficient condition, Denis-Vidal and Joly-Blanchard [25] proposed to verify theidentifiability of uncontrolled and autonomous systems by directly comparing theright hand side function f in Eq. (2.1). Note that here f = f(x(t),θ) does notexplicitly depend on t and u(t) for uncontrolled and autonomous systems. Therefore,the problem becomes whether

f(x,θ1) = f(x,θ2) ⇒ θ1 = θ2

can hold globally or locally. Denis-Vidal and Joly-Blanchard [25] used the followingmodel for quantifying microbial growth [46] to illustrate the basic idea

x(t) =µmbl(t)x(t)

Ks + bl(t)−Kdx(t), (3.11)

l(t) = − µml(t)x(t)

Y (Ks + bl(t)), (3.12)

x(0) = 0, l(0) = 1. (3.13)

Therefore, the right hand side function vector is

f(x, l,θ) =

(µmblxKs+bl −Kdx

− µmlxY (Ks+bl)

),

and from f(x, l,θ1) = f(x, l,θ2) we have

µm1b1lx

Ks1 + b1l−Kd1x =

µm2b2lx

Ks2 + b2l−Kd2x, (3.14)

− µm1 lx

Y1(Ks1 + b1l)= − µm2 lx

Y2(Ks2 + b2l). (3.15)

Solving the two equations above, we have

Kd1 = Kd2 , µm1 = µm2 ,b1b2

=Ks1

Ks2

=Y2

Y1,

which indicates that only parameters (Kd, µm) are identifiable but the rest are not.Although the analytical direct test approach described above is conceptually sim-

ple, it usually requires advanced mathematical skills to obtain analytic solutions andhence is difficult to be applied in practice. If a certain number of state variablesare not measured, such latent variables have to be eliminated first (e.g., by takinghigher order derivatives) in order to use the analytical direct test approach. It maybe necessary to employ computer algebra tools, instead of algebraic manipulations by

Page 9: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 9

hand, for complicated models as suggested by Raksanyi et al. [87]. However, com-puter algebraic computation can easily become unfeasible for complicated nonlinearODE models and Walter et al. [112] illustrated that the conclusions drawn from theanalytical direct test approach can be misleading for certain types of models. Instead,Walter et al. [112] proposed the numerical direct test approach. For a model to beidentifiable, Walter et al. [112] considered the following conditions which should besatisfied in practice,

∃(θ1,θ2) ∈ Rq ×Rq such that y(u,θ1) ≡ y(u,θ2) and ∥ θ1 − θ2 ∥∞> δ,

where δ is a positive cut-off value chosen by the user. Techniques such as the algorithmSIVIA and forward-backward contractor for constraint satisfaction problems (CSP)were employed to find the inner and outer approximations of the solution set S, thatis,

S ⊂ S ⊂ S.

For details of the algorithms for solving CSP (called interval arithmetic), see [53].Walter et al. [112] thought that if S is empty, the model is identifiable; if S is

not empty, then the model is not identifiable. However, the choice of the cut-off valueδ is arbitrary, which seriously restricts the application of the numerical direct testmethod. In the parameter space of some models, there may exist a continuous andflat hypersurface on which the objective function (a function to be minimized for anoptimization problem which evaluates how good a solution is, e.g., the residual sumof squares) has the same minimum value, which suggests the unidentifiability of themodel. Under such circumstances, the numerical direct test approach, however, maystill misleadingly conclude that the model is identifiable. In addition, it is difficultto verify which parameters are identifiable and which are not by using the numericaldirect test method. Therefore, no useful information can be derived from S to help toimprove mathematical models by reparameterizing unidentifiable parameters. Thus,the application of the direct test approach is very limited in practice.

3.3. Differential Algebra. The methods discussed in the previous subsectionsare difficult to apply to general nonlinear systems due to the difficulties in developingsufficient or necessary conditions for system identifiability and solving the resultingequations corresponding to such conditions. Also, rigorous training and advancedmathematical skills are required to use these methods. Is it possible to leave suchtedious algebraic calculations to a computer instead of a human? The idea has mo-tivated the development of methods under the framework of differential algebra [91]and have yielded some promising results [7, 14, 64].

Compared to other methods, the differential algebra approach has the followingadvantages: well-established theories, feasibility to general nonlinear dynamic sys-tems, and availability of several computational algorithms (e.g., [14, 49, 57, 91]) andsoftware packages (e.g., diffalg in Maple c⃝ by Hubert [49], DAISY by Bellu et al.[12]). Theories and algorithms developed in abstract algebra and computer algebraare very helpful to understand differential algebra. For details of abstract algebraand computer algebra, the interested reader is referred to [28] and [72]. For details ofdifferential algebra, the interested reader is referred to [14, 57, 64, 78, 91]. Here weonly review some important concepts, theories and algorithms of differential algebra.

The first important concept is that of a differential polynomial . Here we give thedefinition for general dynamic systems:

Page 10: On Identifiability of Nonlinear Ode Models And

10 H. Miao, X. Xia, A.S. Perelson and H. Wu

Definition 3.2. Differential polynomial: If an expression is constructed fromvariables t, x, u and y, parameter θ = (θc,θt) and constants using only the operationsof addition, subtraction, multiplication, constant positive whole number exponents, andconstant positive whole number derivatives, it is called a differential polynomial.

For example,

y1y22 − 5x1x

22y2 − 3θ1y1y2u1 + θ2y2y

2 (3.16)

is a valid differential polynomial. Note that for the problems considered in this paper,the derivatives in the definition above are with respect to time t only.

Now the dynamic system in Eqs. (2.1) and (2.2) can be rewritten as

x(t)− f(t,x(t),u(t),θ) = 0, (3.17)

y(t)− h(x(t),u(t),θ) = 0. (3.18)

If the left-hand side of Eqs. (3.17) and (3.18) is in the form of a differential polynomialafter necessary algebraic computation or transformation, the structural identifiabilityof this system can be investigated in the differential algebra framework.

Clearly, an infinite number of differential polynomial equations can be formedby adding, scaling, multiplying, and differentiating both sides of Eqs. (3.17) and(3.18). It can be easily proved that the solution to Eqs. (3.17) and (3.18) is also thesolution to all those generated equations. Therefore, the structure identifiability ofEqs. (3.17) and (3.18) can be investigated from those infinite number of generatedequations. Let ℜ{v1, v2, ..., vr} denote the differential polynomial ring with the differ-ential indeterminates v1, v2, ..., vr [91]. For the dynamic systems under consideration,vi ∈ V , i = 1, 2, ..., r can be any component of x, y, u and θ, and ℜ{v1, v2, ..., vr}is the set of the infinite number of generated differential polynomials. As mentionedabove, the derivative on the ring ℜ{v1, v2, ..., vr} is with respect to time t only, sucha ring is called an ordinary differential ring .

Before we introduce more properties of ℜ{v1, v2, ..., vr}, some definitions and con-cepts on differential indeterminates and polynomials need to be described. First, theorder of a differential indeterminate is defined as the order of the derivative of thatindeterminate and the degree of a differential indeterminate is defined as the exponentof that indeterminate. For example, in the first term y1y

22 in Eq. (3.16), the order of

y2 is one and the degree of y2 is two. To compare multiple differential polynomials,ranking needs to be defined [64, 91]:

Definition 3.3. Ranking: A total ordering of all the indeterminates and theirderivatives is called a ranking if

v ≺ v (∀v ∈ V ), v1 ≺ v2 ⇐⇒ v1 ≺ v2 (∀v1, v2 ∈ V )where ≺ means ‘ranks lower than’.

Note that for the same indeterminate v ∈ V , the item with a higher degree rankshigher, e.g., v ≺ v2. The following are two examples of ranking:

u ≺ u ≺ · · · ≺ y ≺ y ≺ · · · ≺ θ ≺ θ ≺ · · · ≺ x ≺ x ≺ · · · , (3.19)

u ≺ y ≺ θ ≺ x ≺ u ≺ y ≺ θ ≺ x ≺ · · · . (3.20)

For a given ranking over a differential polynomial ring ℜ{v1, v2, ..., vr}, the highestranking derivative in a differential polynomial P ∈ ℜ{v1, v2, ..., vr} is called the leaderof P . Therefore, to rank two differential polynomials P1 and P2, the leaders of P1 and

Page 11: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 11

P2 are compared first, then the second highest ranked derivatives are compared if theleaders of P1 and P2 rank the same, and so on. To generalize this ranking conceptto differential polynomials, the concepts of partially reduced and reduced were alsointroduced [64, 91]:

Definition 3.4. Partially reduced: For two differential polynomials P1, P2 ∈ ℜ,let vP1 denote the leader of P1, P2 is said to be partially reduced with respect to P1 ifthere exists no proper derivative of vP1 in P2.

Definition 3.5. Reduced: For two differential polynomials P1, P2 ∈ ℜ, let vP1

denote the leader of P1, P2 is said to be reduced with respect to P1 if P2 is partiallyreduced with respect to P1 and the degree of vP1 in P2 is less than the degree of vP1 inP1.

With the definitions above, an autoreduced set can be introduced as follows:

Definition 3.6. Autoreduced set: A differential polynomial set is said to be anautoreduced set if any two differential polynomials in this set are reduced with respectto each other.

Autoreduced sets can also be ranked [64]. For two autoreduced setsA = {A1, A2, ..., Ar}and B = {B1, B2, ..., Bs}, A ranks lower than B if there exists an integer k, 1 ≤ k ≤min(r, s) such that rank Ai = rank Bi (i = 1, 2, ..., k − 1) and Ak ≺ Bk. ConsiderEqs. (3.17) and (3.18) again, since an infinite number of differential polynomials canbe generated with admissible operations, an infinite number of autoreduced sets canalso be generated. Among these autoreduced sets, the set ranking the lowest is themost important and is called the characteristic set :

Definition 3.7. Characteristic set: Among all the autoreduced sets formed froma finite number of differential polynomials, the set ranking the lowest is called a char-acteristic set.

We now explain the relationship between the characteristic set and structure i-dentifiability. The identifiability problem is to verify whether θ can be uniquelydetermined from differential polynomials with indeterminates u, y and θ only. Ob-viously, there exist an infinite number of sets of differential polynomials that can beemployed to perform the identifiability analysis. However, the characteristic set hasbeen proved to be the ‘best’ set among all such sets [91], where the word ‘best’ meansthe lowest rank. In summary, the characteristic set has the following properties:

(i) Differential polynomials in the characteristic set satisfy Eqs. (3.17) and(3.18);

(ii) Differential polynomials in the characteristic set are in the simplest formpossible;

(iii) Differential polynomials in the characteristic set have the exact informationas in Eqs. (3.17) and (3.18) to verify system identifiability.

A number of algorithms have been developed to find the characteristic set, e.g., theRitt algorithm [91], the Ritt-Kolchin algorithm [57], and the improved Ritt-Kolchinalgorithm [14]. The implementation of such algorithms can be found in the diffgrob2package [67] or the diffalg package [49]. The basic idea of these algorithms is toeliminate the higher ranking variables such as x so that differential polynomials withindeterminates u, y and θ can be obtained via symbolic computations. The keyoperation in the elimination process is called pseudo-division. Before we discuss thedetails of the pseudo-division, more definitions and notations need to be introduced.First, we call the coefficient of the highest power of the leader the initial . In addition,for a differential polynomial P and its leader vP , we call the initial of P the separantof P, denoted by SP . For example, using the ranking (3.19) combined with x1 ≺ x1 ≺

Page 12: On Identifiability of Nonlinear Ode Models And

12 H. Miao, X. Xia, A.S. Perelson and H. Wu

· · · ≺ x2 ≺ x2 ≺ · · · , the initial of (3.16) is (−5x1y2) and the separant is (−10x1x2y2).Considering two differential polynomials P1 and P2, assume the leader of P2 is

vP2 and there exists a proper derivative of vP2 , e.g., v(k)P2

for k ≥ 1 in P1, then P1 can

be partially reduced by P2 as follows [49]. First, take derivatives of P2 up to the kth

orderP2 = SP2 vP2 +R1,P2 = SP2 vP2 +R2,

· · ·P

(k)2 = SP2v

(k)P2

+Rk,where SP2 is the separant of P2, and Ri(i = 1, ..., k) the rest terms. Second, from

the last equation above, v(k)P2

can be expressed in terms of P(k)2 , SP2 and Rk, none of

which contains v(k)P2

. Finally, substitute the expression of v(k)P2

into P1 and then P1 canbe re-written as

SrP2P1 = QP

(k)2 + P ,

where r is an integer, andQ and P are differential polynomials. Furthermore, Q can be

called a pseudo-quotient ; and P contains no proper derivatives of v(k)P2

and can be calleda pseudo-remainder . The procedure described above is called pseudo-division. In thisway, a set of differential polynomials can be reduced to generate an autoreduced set,and eventually, the characteristic set. More details of these computational algorithmscan be found in [14, 49, 57]. However, we notice that the algorithms to find thecharacteristic set are still under development and the existing software packages donot always work well.

Ljung and Glad [64] concluded that each differential polynomial in the character-istic set can be in one of the following three forms if the ranking (3.19) is employed

A1(u,y), ..., Am(u,y),B1(u,y, θ1), B2(u,y, θ1, θ2), ..., Bn(u,y, θ1, ..., θq),

C1(u,y,θ,x), ..., Cl(u,y,θ,x),where A, B and C are differential polynomials and the subindex m, n and l denotethe number of differential polynomials. Ljung and Glad [64] proved a theorem about{B1, B2, ..., Bn} to verify the structural identifiability:

Theorem 3.8. Assume that no separant or initial of {B1, B2, ..., Bn} is zero,(i) If for some Bi, 1 ≤ i ≤ n, in the characteristic set one has Bi = θi, then

the model structure is not identifiable;(ii) If all Bi, 1 ≤ i ≤ n, in the characteristic set are of order zero and degree

one in θi, and there exists non-degenerate solution (y,u,θ∗,x) for some θ∗, the modelstructure is globally identifiable at θ∗;

(iii) If all Bi, 1 ≤ i ≤ n, in the characteristic set are of order zero in θi, andsome Bj is of degree greater than one in θj, and there exists non-degenerate solution(y,u,θ∗,x) for some θ∗, the model structure is locally identifiable at θ∗.

Although the results in Ljung and Glad [64] are for time-invariant parameters,they can be easily extended to time-varying parameter cases by treating θt as statevariables or system inputs [7].

3.4. Implicit Function Theorem. Another approach based on the implicitfunction theorem was proposed by Xia and Moog [119]. For a general introductionof the implicit function theorem, the reader is referred to [73]. The theorem foridentifiability analysis based on the implicit function theorem is given as follows [119]:

Theorem 3.9. Let Φ : Rd+p+q → Rq denote a function of model parameterθ ∈ Rq, system input u ∈ Rn, system output y ∈ Rd and their derivatives, that is,

Page 13: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 13

Φ = Φ(θ,u, u, ...,u(k),y, y, ...,y(k)),

where k is a non-negative integer. Assume that Φ has continuous partial derivativeswith respect to θ. The system structure is said to be locally identifiable at θ∗ if there

exists a point (θ∗,u∗, u∗, ...,u(k)∗ ,y∗, y∗, ...,y

(k)∗ ) ∈ Rd+p+q such that

Φ(θ∗,u∗, u∗, ...,u(k)∗ ,y∗, y∗, ...,y

(k)∗ ) = 0 and

∣∣∣ ∂Φ∂θ∗

∣∣∣ = 0.

We can easily prove this theorem by considering the Taylor expansion of Φ at θ∗Φ ≈ Φ(θ∗) + (θ − θ∗)

∂Φ∂θ∗

;

since Φ(θ∗) = 0 and(

∂Φ∂θ∗

)−1

exists (i.e.,∣∣∣ ∂Φ∂θ∗

∣∣∣ = 0), a unique solution of θ exists

and the system is locally identifiable at θ∗.

Carefully examining this theorem, we find that it is the same as the algebraicalidentifiability definition 2.9. The implicit function theorem method can be employedalone to verify system identifiability, and it can also be used as a supplement tothe differential algebra method. Theorem 3.8 suggests verifying system identifiabilityby examining specific forms of differential polynomials B = {B1, B2, ..., Bn} in thecharacteristic set. However, a more rigorous approach is to verify whether

∣∣∂Φ∂θ

∣∣ = 0as suggested by the implicit function theorem method, where Φ is generated fromB = {B1, B2, ..., Bn}.

Xia and Moog [119] and Jeffrey and Xia [56] proposed a method to generate thefunction Φ by taking higher order derivatives of system output y to eliminate alllatent (unobservable) state variables. For example, consider the following HIV model[81]

dT

dt= λ− ρT − βTV, T (0) = T0, (3.21)

dT ∗

dt= βTV − δT ∗, T ∗(0) = T ∗

0 , (3.22)

dV

dt= NδT ∗ − cV, V (0) = V0, (3.23)

where T is the number of uninfected T cells susceptible to infection (target cells), T ∗

the number of infected T cells, and V the viral load. Using the method of Xia andMoog [119], we can take the 3rd-order derivative of V to obtain

V (3) = (V

V− ρ− βV )[V + (δ + c)V + δcV ] +NλδβV − δcV − (δ + c)V . (3.24)

Therefore, we can define

f = (V

V− ρ− βV )[V + (δ + c)V + δcV ]

+NλδβV − δcV − (δ + c)V − V (3), (3.25)

Φ =[f f (1) f (2) f (3) f (4)

]T, (3.26)

such that Φ = 0 is automatically satisfied. If ∂Φ∂θ = 0, then θ = (β, ρ, ν, µ, η) can be

identified according to the implicit function theorem, where ν = δc, µ = δ + c, and

Page 14: On Identifiability of Nonlinear Ode Models And

14 H. Miao, X. Xia, A.S. Perelson and H. Wu

η = Nλβδ. That is, if

Rank

[∂Φ

∂θ

]= Rank

∂f∂β

∂f∂ρ

∂f∂η

∂f∂µ

∂f∂ν

∂f(1)

∂β∂f(1)

∂ρ∂f(1)

∂η∂f(1)

∂µ∂f(1)

∂ν∂f(2)

∂β∂f(2)

∂ρ∂f(2)

∂η∂f(2)

∂µ∂f(2)

∂ν∂f(3)

∂β∂f(3)

∂ρ∂f(3)

∂η∂f(3)

∂µ∂f(3)

∂ν∂f(4)

∂β∂f(4)

∂ρ∂f(4)

∂η∂f(4)

∂µ∂f(4)

∂ν

= 5, (3.27)

we can identify the five parameters (β, ρ, ν, µ, η) in the model; furthermore, in theoriginal model, N and λ cannot be identified separately. The identification functionΦ involves the 7th order derivative of V which requires at least 8 measurements ofV to evaluate. Such information, a by-product of identifiability analysis, is useful todesign new experiments.

Note that with a high dimensional parameter space, like in the example that weare considering here, the matrix ∂Φ

∂θ can become very complicated to derive and itsrank is difficult to evaluate. For example, one element in the matrix (3.27) is

∂f(4)

∂µ = V (−5)[βV (5)V 6 + 10βV V (3)V 5 + 5βV V (4)V 5 + ρV (5)V 5 + V (6)V 5

−6(V (3))2V 4 − 8V V (4)V 4 − 2V V (5)V 4 + 12V 3V 3 + 44V V V (3)V 3

+9V 2V (4)V 3 − 78V 2V 2V 2 − 32V 3V (3)V 2 + 84V 4V V − 24V 6].Thus, it is not easy to evaluate the rank of the above matrix. To avoid such com-plexity and evaluation of the high order derivatives, an alternative method for for-mulating the identification functions Φ(·) was proposed by Wu et al. [117]. Supposewe have the quantities (V, V , V , V (3)) at five distinct time points t1, · · · , t5. De-

note the values of (V, V , V , V (3)) at t = ti as (Vi, Vi, Vi, V(3)i ) for i = 1, · · · , 5. Let

f1 = f(t1, θ∗, V1, V1, V1, V

(3)1 ), · · · , f5 = f(t5, θ

∗, V5, V5, V5, V(3)5 ), where f is specified

in Eq. (3.25). Then we have

Φ∗ = [f1, f2, f3, f4, f5]T = 0. (3.28)

If

∣∣∣∣∂Φ∗

∂θ

∣∣∣∣ =∣∣∣∣∣∣∣∣∣∣∣

∂f1∂β

∂f1∂ρ

∂f1∂η

∂f1∂µ

∂f1∂ν

∂f2∂β

∂f2∂ρ

∂f2∂η

∂f2∂µ

∂f2∂ν

∂f3∂β

∂f3∂ρ

∂f3∂η

∂f3∂µ

∂f3∂ν

∂f4∂β

∂f4∂ρ

∂f4∂η

∂f4∂µ

∂f4∂ν

∂f5∂β

∂f5∂ρ

∂f5∂η

∂f5∂µ

∂f5∂ν

∣∣∣∣∣∣∣∣∣∣∣= 0, (3.29)

by the implicit function theorem, there exists a unique solution of θ. Assuming thatβ = 0, some algebra shows that the rank of (∂Φ

∂θ ) is equal to the rank of

Σ =

V1(V1 + µV1) V1 + µV1 V1 V 2

1 V1(V−11 V1 − ρ− βV1)− V1

V2(V2 + µV2) V2 + µV2 V2 V 22 V2(V

−12 V2 − ρ− βV2)− V2

V3(V3 + µV3) V3 + µV3 V3 V 23 V3(V

−13 V3 − ρ− βV3)− V3

V4(V4 + µV4) V4 + µV4 V4 V 24 V4(V

−14 V4 − ρ− βV4)− V4

V5(V5 + µV5) V5 + µV5 V5 V 25 V5(V

−15 V5 − ρ− βV5)− V5

. (3.30)

As long as det(Σ) = 0, we have det(∂Φ∗

∂θ ) = 0. Note that in Eq. (3.30), the matrix Σalso involves unknown parameters (µ, ρ, β); thus, to numerically determine Σ’s rank,nominal values of these parameters (i.e., obtained from literature) are needed.

Page 15: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 15

Note that although V (3) is not involved in the matrix Σ, V (3) should exist at anytime point. For evaluating V (3) at one time point, at least four measurements of V areneeded. In order to form the five identification equations, at least eight measurementsof V are necessary. This conclusion is consistent with that of the method proposedby Xia and Moog [119]. Note that this model is more likely to be locally identifiablethan globally identifiable since Σ also contains unknown parameters. Compared to themethod of Xia and Moog [119], the method of Wu et al. [117] is less computationallyintensive and easier to implement since only the lower-order derivatives (the 3rd orlower order derivatives in our case) of V need to be evaluated.

In this section, as a by-product of structural identifiability analysis, we illustratehow to calculate the minimum number of observation points for parameter identifi-ability, which is an important issue in experimental design. Thorough and in depthdiscussions of experimental design for dynamic systems are beyond the scope of thisdocument, so we only briefly discuss the influence of the observation timing on iden-tifiability here. Sontag [100] reported a very general and simple conclusion: for anyODE model with q unknown constant parameters, 2q + 1 experimental observationsare enough for identification of the q parameters if measurements are absolutely accu-rate. Sontag [100] explained that the number 2q+1 is frequently met in geometry anddynamical system theory and it is the embedding dimension in the Whitney’s theo-rem [115, 116] for abstract manifolds or the embedding dimension of q-dimensionalattractors [102]. However, an intuitive explanation for the number 2q is that eachparameter in the righthand side of an ODE model is used to quantify the change ofthe state variables (like a slope) and at least two data points are needed to determinea slope. As to the influence of the observation times on identifiability, the work ofSontag [100] indicated that increasing the number of data points will not help to i-dentify more parameters of a dynamic model once the model enters its steady stateand produces only flat responses all the time. In principle, data points collected tocapture violent nonlinear behavior of the dynamic system will be more informativefor determining parameter values, as suggested in [103].

4. Practical Identifiability Analysis. Before we introduce various techniquesof practical identifiability analysis, note that structural identifiability analysis providesa theoretical ground for practical identifiability analysis. If the structural analysis sug-gests that a model is not theoretically identifiable, the practical analysis is not nec-essary since theoretical unidentifiability must imply practical unidentifiability. Thus,only theoretically identifiable models need further practical identifiability analysis.

Structural identifiability analysis can be done without any actual experimentalobservation, so it is also called prior identifiability analysis. There are two basic as-sumptions upon which structural identifiability analysis heavily rely: model structuresare absolutely accurate and measurements are exact (no measurement errors). Howev-er, these two assumptions are clearly not valid in practice. For instance, in biomedicalresearch, both model uncertainty and measurement error are usually large. Therefore,even when structural identifiability analysis suggests that model parameters can beuniquely identified, the estimates of model parameters may still be unreliable. Thus,it is necessary to evaluate whether structurally identifiable parameters can be reliablyestimated with acceptable accuracy from noisy data. This is so-called practical iden-tifiability analysis or posterior identifiability analysis. It is strongly recommendedto perform both structural and practical identifiability analyses in practice to insurethe reliability of parameter estimation. For the rest of this section, we assume our

Page 16: On Identifiability of Nonlinear Ode Models And

16 H. Miao, X. Xia, A.S. Perelson and H. Wu

measurement or output model to have measurement errors as follows:

y(t) = h(x(t),u(t),θ) + ε(t), (4.1)

where ε(t) is measurement error with mean 0 and variance σ2(t).

4.1. Monte Carlo Simulation. The history of Monte Carlo simulations canbe traced back to the work of Metropolis and Ulam [69]. As implied by its name, thismethod is a sampling technique using random numbers and probability distributions.More specifically, the Monte Carlo simulation method define possible inputs first(e.g., measurement noise level), then randomly generate inputs according to certainprobability distributions (e.g., normal distribution with zero mean), and then use theinputs to do certain calculations (e.g., add random errors to data and fit model to thesimulated noisy data), and finally aggregate individual computation results (e.g., theaverage error in parameter estimates). It is not only useful for practical identifiabilityanalysis but also helpful for experimental design. Monte Carlo simulation is verypopular and widely used to assess the performance of statistical estimation methodsin the statistical literature.

Once parameters or a subset of parameters of a model are determined to betheoretically (structurally) identifiable, one can use Monte Carlo simulations to eval-uate whether the theoretically identifiable parameters can be reliably estimated withacceptable accuracy from noisy data. Obviously, in order to evaluate the practical(statistical) identifiability, statistical estimation methods, such as the least squaresapproach, need to be readily available. However, statistical parameter estimationfor nonlinear ODE models is beyond the scope of this review paper and will not bereviewed here.

Monte Carlo simulations allow us to simulate various scenarios with differentnumbers of observations at different levels of noise or measurement error for differentexperimental designs although such designs may not be feasible for practical experi-ments. The simulated data can be used to evaluate whether model parameters or asubset of parameters can be reliably estimated under different conditions. In general,a Monte Carlo simulation procedure can be outlined as follows:

(i) Determine the nominal parameter values θ0 for simulation studies, whichcan be obtained by fitting the model to experimental data if available. Otherwise itcan be obtained from the literature or other resources.

(ii) Use the nominal parameter values to numerically solve the ODE model toget the solution of the output or measurement variables at the experimental designtime points.

(iii) Generate N sets (e.g., 1000) of simulated data from the output or measure-ment model (4.1) with a given measurement error level.

(iv) Fit the ODE model to each of the N simulated data sets to obtain parameterestimate θi, i = 1, 2, ..., N .

(v) Calculate the average relative estimation error (ARE) for each element of θas

ARE = 100%× 1

N

N∑i=1

∣∣∣θ(k)0 − θ(k)i

∣∣∣∣∣∣θ(k)0

∣∣∣ , (4.2)

where θ(k)0 is the k-th element of θ0 and θ

(k)i is the k-th element of θi.

Page 17: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 17

The ARE can be used to assess whether each of the parameter estimates is ac-ceptable or not. For a very small measurement error, the parameter estimates shouldbe close to the true values and the ARE should be close to 0. When the measurementerror increases, the ARE of the parameter estimates will also increase. However, theARE for some of the parameter estimates may increase significantly and some othersmay just increase a little. However, for a reasonable or practical level of measure-ment error, if the ARE of a parameter estimate is unacceptably high, we claim thatthis parameter is not practically or statistically identifiable. In practical applications,some parameters may not be sensitive to measurement errors and can always be wellestimated, some other parameters may be quite sensitive to measurement errors andtheir AREs are large even for a small measurement error, and at the same time, someparameters may be in the middle ground [70]. In addition, there is no clear cut ruleon how high the ARE need to be before they are claimed “unacceptable” for a par-ticular problem. Thus, the practical identifiability relies on the underlying problemand judgement of the investigators. Also notice that various statistical estimationapproaches can be employed to obtain the parameter estimates, and the ARE maydepend on the estimation methods.

Monte Carlo simulations can also be used to design better practical experiments.Different designs for different sample sizes under different conditions can be evaluatedusing the AREs. The best design and the necessary sample size can be determinedbased on the Monte Carlo simulation results. We will illustrate the application of thismethod in § 6.1.

4.2. Correlation Matrix. Although the Monte Carlo simulation approach iseasy to understand and simple to implement, the associated computational cost ishigh since a large number of model fits to data need to be performed. Rodriguez-Fernandez et al. [92, 93] proposed an alternative approach for practical identifiabilityanalysis of ODE models by examining the correlations between model parameters.This requires much less computation and is relatively simple if measurement errorsfollow an identical and independent distribution (i.i.d.).

The idea behind this approach is straightforward. Assume that the parameterestimate θ = [θ1, θ2, · · · , θq]T has been obtained after fitting a model to experimentaldata. The correlation matrix of the parameter estimates can then be calculated basedon the Fisher information matrix (FIM) [33, 109] in the following form:

R =

r11(θ1, θ1) r12(θ1, θ2) · · · r1q(θ1, θq)

r21(θ2, θ1) r22(θ2, θ2) · · · r1q(θ2, θq)...

rq1(θq, θ1) rq2(θq, θ2) · · · rqq(θq, θq)

, (4.3)

where rij (i, j = 1, 2, ..., q and −1 ≤ rij ≤ 1) is the correlation coefficient between

parameter estimates θi and θj . If there exists a strong positive correlation between

parameter estimates θi and θj , that is, the correlation coefficient rij is close to 1,parameters θi and θj are said to be practically undistinguishable. A strong correlationbetween two parameters indicates that one parameter strongly depends on anotherparameter and these two parameters cannot be separately estimated.

A derivation of the expression for the correlation matrix was provided by Rodriguez-Fernandez et al. [92]. For simplicity, the measurement errors were assumed to beuncorrelated and follow an identical normal distribution with mean zero, that is,

Page 18: On Identifiability of Nonlinear Ode Models And

18 H. Miao, X. Xia, A.S. Perelson and H. Wu

N(0, σ2). In this case, for a general dynamic system (2.1) and (2.2), the Fisher infor-mation matrix can be given as

FIM =N∑i=1

(∂yi

∂θ

)T

V−1

(∂yi

∂θ

), (4.4)

where the subscript i denotes the ith time point of experimental observation, N thetotal number of observations, yi the model approximation of observation, θ the modelparameter estimate, and V a known positive definite matrix of weights on variances.It can be proved that the covariance matrix C is equal to the inverse of FIM accordingto the Cramer-Rao theorem [89], that is

C = FIM−1. (4.5)

Finally, the element rij of the correlation matrix can be defined as

rij =Cij√CiiCjj

, i = j, (4.6)

rij = 1, i = j. (4.7)

Guedj et al. [40] tackled the practical identifiability problem of HIV dynamicmodels. They developed their approach under the framework of maximum likelihoodestimation instead of least squares estimation, but their results are still based on theFisher information matrix and the idea is similar to that in [92].

A limitation of the correlation matrix approach is that it requires not only theparameters but also their correlation matrix to be reliably estimated. This may be aproblem for an model with most parameters unidentifiable since the correlation ma-trix estimate may strongly depend on the parameter estimates. If any two parametersare not distinguishable, the parameter estimates and their correlation matrix estimatemay be poor. In addition, the correlation matrix approach only allows you to checkwhether any pair of parameters are distinguishable or not; to evaluate correlationsbetween more than two parameters, the sensitivity-based identifiability analysis tech-niques (e.g., eigen-decomposition of the sensitivity matrix) should be considered, asdescribed in the next section.

5. Sensitivity-Based Identifiability Analysis. Sensitivity analysis itself is arich topic. The interested reader is referred to the comprehensive survey by Saltelli etal. [94] and Cacuci [13]. Sensitivity analysis (SA) is often used to assess the variationof system output induced by different input factors including model parameters. Thesensitivity analysis idea can also be used to evaluate the identifiability of unknownparameters.

Sensitivity-based identifiability analysis is similar to the structural analysis ap-proach in the sense that both approaches do not require actual experimental data(although the sensitivity-based method could require the number and locations ofmeasurement time points, see details below), and both approaches assume that mea-surements are precise without error. However, the sensitivity-based method does notdirectly use the model structure information, which is a critical difference betweenthe structural and the sensitivity-based approaches. The sensitivity-based methodis similar to the practical analysis approach in the sense that both methods requirepre-specified parameter values (either nominal or actual estimates), and both need to

Page 19: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 19

know the number and locations of measurement time points. However, the sensitivity-based method is different from the practical analysis approach in the sense that thesensitivity-based method does not take measurement error into account. Thus, thesensitivity-based method is a technique between the structural (theoretical) identifi-ability and practical identifiability analysis. We review such methods in this section.

A nominal parameter value is required for the sensitivity-based approach. Thus,parameter identifiability is evaluated with respect to a specific point in the parameterspace by sensitivity-based methods. For this reason, the concept of at-a-point iden-tifiability was introduced by Ljung and Glad [64] and Quaiser and Monnigmann [86]as follows,

Definition 5.1. Globally at-a-point identifiable: Let θ∗ denote a fixed point inthe parameter space Θ. A system is said to be globally at-a-point identifiable if for anyadmissible input u(t) and any parameter vector θ ∈ Θ, y(u,θ) = y(u,θ∗) impliesθ = θ∗.

Definition 5.2. Locally at-a-point identifiable: Let θ∗ denote a fixed point inthe parameter space Θ. A system is said to be locally at-a-point identifiable if for anyadmissible input u(t) and any parameter vector θ within an open neighborhood of θ∗,y(u,θ) = y(u,θ∗) implies θ = θ∗.

The sensitivity-based identifiability analysis techniques reviewed in this sectiononly examine at-a-point identifiability. The sensitivity of measurable system respons-es with respect to parameter values is used to assess the identifiability of unknownparameters by these methods. More specifically, assume that the locations and thenumber of time points at which the system responses or state variables will be mea-sured have been given, denoted by t1 ≤ t2 ≤ · · · ≤ tN , then the sensitivity coefficientat each time point tk (k = 1, 2, ..., N) for a given nominal parameter vector θ∗ isdefined as

sij(tk) =∂yi(tk,θ

∗)

∂θj, (5.1)

where yi(i = 1, 2, ..., d) denotes the ith component of y (y ∈ Rd) and θj the jth(j =1, 2, ..., q) component of θ (θ ∈ Rq). Thus, the sensitivity matrix for all time pointsis defined as

SdN×q =

s11(t1) · · · s1q(t1)

· · ·. . . · · ·

sd1(t1) · · · sdq(t1)...

......

s11(tN ) · · · s1q(tN )

· · ·. . . · · ·

sd1(tN ) · · · sdq(tN )

. (5.2)

A number of identifiability analysis techniques have been developed based on thissensitivity matrix. Simply speaking, the larger the sensitivity coefficients are, themore notable the system responses or measurable state variables are with respectto the changes of parameters. In that sense, a parameter is likely to be identifiableif the system output is highly sensitive to a small perturbation of this parameter;otherwise, the parameter is likely to be unidentifiable. In addition, if there exists a

Page 20: On Identifiability of Nonlinear Ode Models And

20 H. Miao, X. Xia, A.S. Perelson and H. Wu

strong correlation between any two parameters, these two parameters are very likelyto be indistinguishable from each other. Such parameter dependence can also beevaluated by examining the dependence of the sensitivity matrix columns. We reviewfour typical methods along this line in detail: the correlation method [51, 93, 122],the principle component analysis (PCA) method [24, 34, 54], the orthogonal method[35, 120, 121], and the eigenvalue method [85, 86, 96, 108].

5.1. Correlation Method. The correlation method was first proposed by Jacquezand Greif [51]. The method was originally developed to study identifiability for linearcompartment models and the derivation was given for a single output system only(that is, y ∈ R). However, this method is not limited to linear models and singleoutput systems.

Consider the first order Taylor expansion of the system output near the pre-specified nominal parameter vector θ∗

yk(θ) = y(x(tk),u(tk),θ)

≈ y(x(tk),u(tk),θ∗) +

∂y(x(tk),u(tk),θ)

∂θ

∣∣∣∣θ=θ∗

· (θ − θ∗), (5.3)

where k = 1, 2, ..., N denotes the index of the measurement time points. Let rk denotethe measurement at tk without errors and △θ = θ − θ∗, then the residual sum ofsquares between the exact measurements and the linear approximation is

RSS(△θ) =N∑

k=1

[rk − yk(θ

∗)− ∂y(x(tk),u(tk),θ)

∂θ

∣∣∣∣θ=θ∗

· △θ

]2

=N∑

k=1

[∂y(x(tk),u(tk),θ)

∂θ

∣∣∣∣θ=θ∗

· △θ

]2, (5.4)

where rk − yk(θ∗) = 0 based on our assumptions. Finally, we can rewrite Eq. (5.4)

in terms of the sensitivity matrix,

RSS(△θ) = (S△θ)T · S△θ, (5.5)

where S is the sensitivity matrix defined in (5.2). Obviously, the minimum ofRSS(△θ)is reached at STS ·△θ = 0. If STS is of full rank, the unique solution of STS ·△θ = 0is θ = θ∗, which indicates that the model parameters θ are locally identifiable atθ∗. If STS is singular, then there exists at least one non-trivial solution θ = θ∗ suchthat the model parameters are not identifiable at θ∗. Two important issues should benoticed: first, a similar expression can be derived under the framework of maximumlikelihood estimation [40], so the derivation is not limited to ordinary least squares;secondly, only local identifiability can be inferred based on the rank of STS since thelinear approximation Eq. (5.3) is used [86].

It is also desirable to determine which parameters are not identifiable if STS is notof full rank. For this purpose, the correlations between parameters can be calculatedbased on the sensitivity matrix. More specifically, if we examine the columns of thesensitivity matrix defined in Eq. (5.2), it is clear that each column is the sensitivityof the system responses at all time points with respect to one particular parameter.Thus, the sample correlation of two columns is an estimate of the correlation betweentwo parameters that can be calculated as

corr(S·i,S·j) =cov(S·i,S·j)

σ(S·i)σ(S·j), (5.6)

Page 21: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 21

where S·i (or S·j ) denotes the ith (or jth) column of the sensitivity matrix S,cov(S·i,S·j) the sample covariance between S·i and S·j , and σ(S·i) and σ(S·j) thesample standard deviations of S·i and S·j , respectively. If the calculated correlationcoefficient between any two parameters is close to one, these two parameters are notdistinguishable. However, such a decision always involves a pair of parameters. Is itpossible to determine which one in this pair is more problematic and should be fixedor removed from the model? Quaiser and Monnigmann [86] proposed the concept oftotal correlation for this question,

ctoti =

q∑j=1,j =i

|corr(S·i,S·j)| · I(|corr(S·i,S·j)| ≥ 1− δ), (5.7)

where I denotes the indicator function, and δ ∈ (0, 1) the cut-off value specified bythe user. The parameter with the highest total correlation is the first candidate to befixed or removed from the model.

If we compare the correlation method with the correlation matrix method forpractical identifiability analysis described in the previous section [92], we can findthat the Fisher information matrix and the sensitivity matrix are somehow similar;however, the ways of calculating correlations are different.

5.2. Tuning Importance Method and Principle Component Analysis.This category of methods have been developed to reduce model complexity by dis-carding nonsignificant parameters, such as the tuning importance method [24, 95, 105]and principle component analysis(PCA) [24, 34, 54]. More specifically, both methodsrank all parameters first, and then these parameters are determined as identifiable orunidentifiable according to their ranks.

One important and interesting feature of the tuning importance and PCA meth-ods is that they are based on the normalized sensitivity matrix, which is different fromthe sensitivity matrix defined in Eq. (5.2). To construct the normalized sensitivitymatrix, the dimensionless sensitivity coefficient was defined as follows [24, 30],

sik,j =θjyi

· ∂yi(tk,θ)∂θj

=∂ ln yi(tk,θ)

∂ ln θj, (5.8)

where i ∈ {1, 2, ..., d} denotes the index of system outputs, j ∈ {1, 2, ..., q} the indexof parameters and k ∈ {1, 2, ..., N} the index of measurement time points. Then thenormalized sensitivity matrix for each yi is defined as

SiN×q =

si1,1 · · · si1,q...

. . ....

siN,1 · · · siN,q

. (5.9)

For the tuning importance method, the following objective function was intro-duced by Seigneur et al. [95] and Turanyi [105],

e(△θj) =

N∑k=1

d∑i=1

[yi(tk,θ−j , θj +△θj)− yi(tk,θ)

yi(tk,θ)

]2, (5.10)

where θ−j denotes the parameter vector with the jth component removed. By follow-ing the same procedure as the correlation method in the last subsection, the overall

Page 22: On Identifiability of Nonlinear Ode Models And

22 H. Miao, X. Xia, A.S. Perelson and H. Wu

sensitivity can be obtained and expressed in terms of the dimensionless sensitivitycoefficients

os(θj) =∂e

∂ ln θj=

N∑k=1

d∑i=1

[∂ ln yi(tk,θ)

∂ ln θj

]2=

N∑k=1

d∑i=1

(sik,j)2

. (5.11)

Thus, the larger the overall sensitivity of one parameter, the more sensitive the systemresponse is with respect to small perturbations of this parameter. Since all the pa-rameters can be ranked according to their overall sensitivities, the parameters rankingthe lowest are candidates to be unidentifiable and to be discarded.

For the PCA method, the eigenvalues and eigenvectors of the matrix SiTSi arecalculated to provide the information for model reduction. Let λi

j denote the eigen-values which are ordered non-decreasingly

|λi1| ≤ |λi

2| ≤ · · · ≤ |λiq|. (5.12)

Also, let the corresponding eigenvectors be denoted by

Γi = (γi1,γ

i2, · · · ,γi

q) =

γi1,1 · · · γi

1,q...

. . ....

γiq,1 · · · γi

q,q

.

Three strategies examining the eigenvalues and eigenvectors were proposed to rankall the parameters by Jolliffe [54]:

(i) Starting with the eigenvector corresponding to the smallest absolute eigen-value, loop over each eigenvector to locate the component with the maximum absolutevalue and the corresponding parameter at the maximum component location is markedas unidentifiable and to be removed if it has not be marked before. This procedure issummarized as follows:

m1 = arg

(max1≤j≤q

|γj,1|),

ml = arg

max1≤j≤q

j =m1,··· ,ml−1

|γj,l|

, for l > 1.(5.13)

(ii) Starting with the eigenvector corresponding to the smallest absolute eigen-value again, loop over each row of matrix Γi and calculate the sum of squares of allcomponents in each row. The parameter corresponding to the location of the row withthe largest sum of squares is determined as unidentifiable for removal. This procedureis summarized as follows:

m1 = arg

(max1≤j≤q

q∑h=1

γ2j,h

),

ml = arg

max1≤j≤q

j =m1,··· ,ml−1

q∑h=1

γ2j,h

, for l > 1.

(5.14)

(iii) Starting with the eigenvector corresponding to the largest absolute eigen-value, loop over each eigenvector component to locate the largest one and mark theparameter if it is not marked before. The marked parameter is not selected as an

Page 23: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 23

unidentifiable parameter immediately; instead, a rank is assigned to this parame-ter. Eventually, all parameters are ranked and the parameters ranking the lowest arethought unidentifiable. This procedure is summarized as follows:

m0 = arg

(max1≤j≤q

|γj,q|),

ml = arg

max1≤j≤q

j =m0,··· ,ml−1

|γj,q−l|

, for l > 0.(5.15)

Froemel [34] proposed a simple strategy to integrate the rankings from all thethree strategies described above. Further details are given in [34] and [86].

5.3. Orthogonal Method. The orthogonal method was proposed by Yao et al.[120]. The basic idea of this approach is to examine the (nearly) linear dependenciesof columns of the sensitivity matrix S defined in Eq. (5.2). Thus, both the sensitivityof system response with respect to parameter values and the dependency betweenparameters regarding their effects on the system responses can be simultaneouslyevaluated to determine a set of identifiable parameters.

Unlike the correlation method, the orthogonal method does not calculate thecorrelation between different columns of S. Instead, the perpendicular distance of onecolumn to the vector space spanned by other columns is calculated as a measurementof the linear dependency. It is an iterative procedure. More specifically, at the firstiteration, the column of S with the largest sum of squares is removed from S andselected as the first element of an empty set SI . At the (j + 1)th (j ∈ {1, · · · q − 1})iteration, j columns have been removed from S and selected into SI , and a vectorspace spanned by all the columns in SI is denoted by VSI

. For column Sh still in S,the orthogonal projection Sproj

h of this column on the vector space VSI is calculatedand the perpendicular is then obtained as

S⊥h = Sh − Sproj

h . (5.16)

In the work of Yao et al. [120], the norm ∥S⊥h ∥ was proposed to be the measurement

of nearly linear dependency between the vector Sh and the vector space VSIsince the

shorter the distance is, the larger the dependence. For a norm ∥S⊥h ∥ which is nearly

zero, the corresponding column Sh is thought to be nearly linearly dependent andthus is not identifiable. At the first iteration, the column of S with the largest norm isselected into SI . However, another reasonable alternative criterion is to consider the

angle between Sh and Sprojh (that is, arccos

(Sh·Sproj

h

∥Sh∥·∥Sprojh ∥

)) since this criterion can

select the best candidate even if the norms of different columns are orders of magnitudedifferent. Finally, at each iteration, a pre-specified cut-off value δ will be used for theperpendicular distance or the angle of all columns in S. Once the distance or angleof one column is smaller than δ, this column is thought to be linearly dependent onVSI

; therefore, the column with the largest ∥S⊥h ∥ is removed from S and selected

into the set SI . This procedure is repeated until S becomes empty. The vectorsin SI determine which parameters are identifiable, which is the primary goal of theorthogonal method (i.e., find the identifiable rather than unidentifiable parameters).

Since the cut-off value δ in this method is an arbitrary value specified by users,the number of unidentifiable parameters strongly depends on the selection of δ. Dueto this problem, Quaiser and Monnigmann [86] proposed to rank all the parametersbased on the values of norms or angles instead of simply dividing them into identifiableor unidentifiable groups.

Page 24: On Identifiability of Nonlinear Ode Models And

24 H. Miao, X. Xia, A.S. Perelson and H. Wu

5.4. Eigenvalue Method. The eigenvalue method was first proposed by Vajdaet al. [108] and then further developed by Quaiser et al. [85], Schittkowski [96], andQuaiser and Monnigmann [86]. This approach is based on the properties of eigenvaluesand eigenvectors of STS, where S is the sensitivity matrix defined in Eq. (5.2). Toillustrate this method, consider the residual sum of squares between system outputsand experimental measurements

RSS(θ) =N∑

k=1

[rk − yk(θ)]2, (5.17)

and let λ1 ≤ λ2 ≤ · · · ≤ λq denote the eigenvalues of STS in a non-decreasing order

and (γ1,γ2, · · · ,γq) the corresponding normalized eigenvectors. Note that since STSis symmetric and positive semi-definite, all its eigenvalues are real and non-negative.Given a nominal parameter vector θ∗, the Taylor expansion of RSS at θ∗ along aneigenvector is approximately

RSS(θ∗ + αγj) ≈ RSS(θ∗) +∇RSS(θ∗) · αγj +1

2α2γT

j · STS · γj , (5.18)

where α is an arbitrary small constant and ∇RSS(θ∗) is the gradient of RSS at θ∗.Although ∇RSS(θ∗) is not necessarily zero if rk is not an exact measurement, sinceθ∗ is a nominal parameter vector that may not minimize RSS, the second order term12α

2γTj · STS · γj can become zero if the eigenvalue λj corresponding to γj is equal

to zero since STS · γj = λjγj and γTj γj = 1. That is, along the direction of γj with

λj = 0, the change of RSS is expected to be nearly zero. The selection criterion forunidentifiable parameters is given by

m = arg

(max1≤h≤q

(|γj,h|)). (5.19)

In practice, λj is usually not exactly zero, therefore, a cut-off value δ needs to bespecified. For a detailed implementation algorithm, the interested reader is referredto [86].

The four sensitivity-based identifiability analysis methods described above werealso reviewed and compared in Quaiser and Monnigmann [86]. In general, all fourapproaches are applicable to general ODE models; however, the eigenvalue methodand the orthogonal method are better designed to globally evaluate and compare theinfluences of parameter values on system outputs such that these two methods out-perform the correlation method and the principle component analysis method. Inaddition, both the eigenvalue method and the orthogonal method are easy to imple-ment. Note that, the method proposed in [88] made the assumption that the nominalparameter vector θ∗ minimizes the objective function, which should be interpreted asrk = yk(θ

∗) as in (5.4) for the correlation method. In addition, it should be men-tioned that in practice, if the sensitivity matrix is of full rank but with eigenvaluesof different orders of magnitude, the parameters corresponding to the smallest eigen-values are theoretically identifiable but likely to be practically unidentifiable. Undersuch a circumstance, the sensitivity-based approaches are still useful in the sense ofdetermining practically unidentifiable parameters.

Finally, it is also interesting to combine the sensitivity-based approaches with thepractical identifiability methods introduced in Section 4 to study the identifiabilityof a dynamic system. Notice that the sensitivity-based approaches do not require

Page 25: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 25

statistical estimation of unknown parameters, which can be done before the practicalidentifiability analysis.

6. Application Examples. In this section, we illustrate the applications ofboth structural and practical identifiability analysis techniques through examples inmodeling viral dynamics. We summarize the identifiability analysis results for popularmodels of HIV infection, influenza infection, and hepatitis B and C virus dynamics.

6.1. HIV Model with Constant Parameters. Miao et al. [70] proposed thefollowing model to describe a growth competition assay to quantify HIV replicationfitness:

dT

dt= (ρ− kmTm − kwTw − kRTmw)T, (6.1)

dTm

dt= (ρm + kmT − qmTw)Tm + 0.25kRTmwT, (6.2)

dTw

dt= (ρw + kwT − qwTm)Tw + 0.25kRTmwT, (6.3)

dTmw

dt= (ρmw + 0.5kRT )Tmw + (qm + qw)TmTw, (6.4)

where T , Tm, Tw and Tmw are the numbers of uninfected cells, cells infected by mutantviruses, cells infected by wild-type viruses, and cells infected by both mutant and wild-type viruses (dual-infection). Let (λ, λm, λw, λmw) represent the proliferation ratesof T , Tm, Tw and Tmw and (δ, δm, δw, δmw) the death rates of T , Tm, Tw and Tmw,respectively. Then ρ = λ − δ, ρm = λm − δm, ρw = λw − δw and ρmw = λmw − δmw

are the net growth rates of T , Tm, Tw and Tmw, which are the differences betweenthe corresponding proliferation rates and death rates. Parameters (km, kw, kR) areinfection rates of mutant virus, wild-type virus and recombinant virus respectively,and qm and qw are dual infection rates.

For this example, the implicit function method of Xia and Moog [119] is employedto investigate the structural identifiability. Since all state variables (T, Tm, Tw, Tmw)are experimentally measurable, the outputs of the system are

y1 = T, y2 = Tm, y3 = Tw, y4 = Tmw. (6.5)

By taking derivatives of one of the four equations in this HIV viral fitness model, thestructural identifiability of the model can be evaluated. To demonstrate this, here westart with the first equation,

y1 = ρy1 − kmy1y2 − kwy1y3 − kRy1y4. (6.6)

By taking higher orders (up to the 4th order) of derivatives on both sides of Eq. (6.6),we get

y1 = ρy1 − km(y1y2)(1) − kw(y1y3)

(1) − kR(y1y4)(1), (6.7)

y(3)1 = ρy1 − km(y1y2)

(2) − kw(y1y3)(2) − kR(y1y4)

(2), (6.8)

y(4)1 = ρy

(3)1 − km(y1y2)

(3) − kw(y1y3)(3) − kR(y1y4)

(3), (6.9)

When Eqs. (6.7)-(6.9) are written in matrix form, we find that the parameters are

Page 26: On Identifiability of Nonlinear Ode Models And

26 H. Miao, X. Xia, A.S. Perelson and H. Wu

identifiable if

Rank

y1 −y1y2 −y1y3 −y1y4y1 −(y1y2)

(1) −(y1y3)(1) −(y1y4)

(1)

y1 −(y1y2)(2) −(y1y3)

(2) −(y1y4)(2)

y(3)1 −(y1y2)

(3) −(y1y3)(3) −(y1y4)

(3)

= 4. (6.10)

Note that the rank of this matrix can be evaluated numerically if the analytical formis not available, and nominal parameter values are not needed for this case since thematrix above involves no parameters. Since the left hand side of Eq. (6.9) has aderivative of order 4, at least five measurements of y1 = T are needed to evaluate

y(4)1 , and at least four measurements of y2 = Tm, y3 = Tw, and y4 = Tmw are neededto evaluate their derivatives of order 3. Since km and kR can be identified from Eqs.(6.7)-(6.9) if Eq. (6.10) holds, Eq. (6.2) can be rewritten as

y2 − kmy1y2 − 0.25kRy1y4 = ρmy2 − qmy2y3. (6.11)

Similarly, by taking the higher derivatives of Eq. (6.11), the parameters (ρm, qm) areidentifiable if

Rank

[y2 −y2y3y2 −(y2y3)

(1)

]= 2. (6.12)

By the same token, the parameters (ρw, qw) are identifiable if

Rank

[y3 −y2y3y3 −(y2y3)

(1)

]= 2. (6.13)

Finally, ρmw is identifiable if

Rank [y4] = 1. (6.14)

In summary, all parameters (ρ, ρm, ρw, ρmw, km, kw, kR, qm, qw) are structurallyidentifiable if at least five measurements of y1 = T and four measurements of y2 = Tm,y3 = Tw, and y4 = Tmw are available and if all coefficient matrices (6.10), (6.12), (6.13)and (6.14) are of full rank at least for some local time points.

Monte Carlo simulations were also performed to evaluate the practical identifia-bility of this HIV viral fitness model by Miao et al. [70]. The AREs of all 9 parametersfor three measurement error levels (0%, 5%, and 30%) are duplicated in Table 6.1.Note that in this simulation study, we assumed that there were 1000 replicates ofdata at each time point although this may not be feasible in practical experiments.But this will help us to evaluate whether the unknown parameters are practicallyidentifiable when the sample size of the data (with noise) is large enough. We can seethat when there is no measurement error (σ = 0%), all the 9 parameters can be wellidentified (the maximum ARE is 0.4%), which confirms our theoretical identifiabilityanalysis results. This also indicates that the parameter estimation method is goodand the parameter estimates converges to the true parameter values when the samplesize is large enough and the measurement error is small enough. However, when themeasurement error increases to 5% and 30%, the ARE of parameter ρmw rapidly in-creases to 556% and 2062%, respectively. The ARE of ρw also increases to 39% and201%, while the ARE of kR increases to 28% and 106%, respectively. The AREs ofparameters (qw, ρm) are reasonable for the case of small measurement error (σ = 5%),

Page 27: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 27

Table 6.1Practical identifiability analysis by Monte Carlo simulations. The average relative error (ARE)

is calculated based on 1000 simulation runs; 1000 replicates at each time point are generated foreach simulated data set (Table 6 in Miao et al. [70]).

Error ρ ρm ρw ρmw km kw kR qm qwLevel(%) (%) (%) (%) (%) (%) (%) (%) (%) (%)

0 0.002 0.017 0.034 0.400 0.003 0.004 0.026 0.003 0.0095 1.1 10.8 39.0 555.7 2.9 4.6 28.3 4.2 12.030 6.5 49.1 201.4 2062 12.7 23.1 106 21.3 59.1

Table 6.2Practical identifiability analysis using Monte Carlo simulations based on 1000 simulation runs

with the measurement error level σ = 1.5% (Table 7 in Miao et al. [70]).

Time Replicate ρ ρm ρw ρmw km kw kR qm qwPoint (%) (%) (%) (%) (%) (%) (%) (%) (%)5 3 6.28 52.7 187 2130 13.5 22.0 108 21.1 54.95 6 3.91 34.7 142 1402 8.99 16.8 77.5 14.4 37.55 9 3.15 30.0 133 1396 8.18 15.3 73.9 13.7 32.35 100 1.01 8.11 40.1 459.8 2.32 4.70 25.2 3.72 10.39 3 4.10 37.9 146 1786 10.1 16.8 88.6 16.1 36.79 6 3.10 27.9 118 1301 7.2 13.1 66.7 10.3 29.59 9 2.72 21.5 85.6 1190 5.84 10.1 61.5 9.24 26.19 100 0.76 6.94 28.74 410 1.79 3.43 22.26 2.98 7.71

but increase to 49% and 59% for the large measurement error case (σ = 30%), re-spectively. The AREs for other four parameters (ρ, km, kw, qm) are reasonable for allcases.

To further investigate the practical identifiability of unknown parameters un-der practical experimental conditions, we performed more simulations for differentnumbers of time points and different numbers of replicates at each time point. Thesimulation results are given in Table 6.2. One can see that the ARE of parameterρmw ranges from 410% to 2130%. This, combined with the results in Table 6.1, in-dicates that the ρmw is practically unidentifiable. Considering the practical case of 9time points and 9 replicates for each time point, the ARE of parameter ρw is 86% ,which indicate that it may be difficult to accurately identify the parameter ρw unlessthe sample size is unrealistically large (say, 100 replicates for each time point). Forparameter kR, the AREs are also large (ranging from 62% to 108%) for practicalcases (the number of replicates is 3, 6, or 9). For parameters (ρm, qw), the AREs arereasonable (ranging from 22% to 38%) for most reasonable sample sizes, thus (ρm, qw)can be considered as reasonably identifiable. The parameters (ρ, km, kw, qm) are verywell identified (the AREs ranging from 3% to 22%) in all cases.

6.2. HIV Model with Constant and Time Varying Parameters. In thissection, we consider another dynamic system that is widely used to describe HIVdynamics in HIV-infected patients with antiretroviral treatment [17, 48, 82]:

d

dtTU (t) = λ− ρTU (t)− η(t)TU (t)V (t), (6.15)

d

dtTI(t) = η(t)TU (t)V (t)− δTI(t), (6.16)

d

dtV (t) = NδTI(t)− cV (t), (6.17)

Page 28: On Identifiability of Nonlinear Ode Models And

28 H. Miao, X. Xia, A.S. Perelson and H. Wu

where TU is the concentration of uninfected target cells, TI the concentration ofinfected cells, V (t) the viral load, λ the source rate of uninfected T cells, ρ the deathrate of uninfected T cells, η(t) the time-varying infection rate which is a functionantiviral treatment efficacy, δ the death rate of infected cells, c the clearance rate offree virions, and N the average number of virions produced by a single infected cellover its lifetime. TU (t), TI(t) and V (t) are state variables, and (λ, ρ,N, δ, c, η(t))T areunknown dynamic parameters.

The differential algebra approach for structural identifiability analysis (Section3.3) requires one to eliminate the latent (unobservable) state variables from the dy-namic equation in order to evaluate the identifiability. The concept of ranking isintroduced such that computer algorithms can be designed to eliminate variables ortheir derivatives with higher rank. For notation simplicity, let x1, x2 and x3 denoteTU , TI and V , respectively. In the dynamic model (6.15)-(6.17), we can measure viralload (x3 = V ) and total CD4+ T cell counts (x1 + x2 = TU + TI). Let y1 and y2denote the measurable variables x1 + x2 and x3, respectively. We adopt the ranking

η ≺ y2 ≺ y2 ≺ θ ≺ x3 ≺ x2 ≺ x1, (6.18)

where θ = [λ, ρ,N, δ, c]T is the vector of constant unknown parameters. We caneliminate x1, x2 and x3 using the ranking (6.18) to obtain

y1 + (ρ+ δ)y1 + δρy1 − δλ+ η(t)y2(y1 + δy1 − λ) = 0, (6.19)

y2 + (δ + c)y2 + δcy2 − η(t)y2(Nδy1 − y2 − cy2) = 0. (6.20)

Note η(t) can be expressed in terms of measurable state variables and other unknownconstant parameters either from Eq. (6.19) as

η(t) =y1 + (ρ+ δ)y1 + δρy1 − δλ

−y2(y1 + δy1 − λ), (6.21)

or from Eq. (6.20) as

η(t) =y2 + (δ + c)y2 + δcy2y2(Nδy1 − y2 − cy2)

. (6.22)

Thus, the time-varying parameter η(t) is identifiable if all the constant parametersare identifiable.

To verify the identifiability of constant parameters θ, Eq. (6.21) and Eq. (6.22)can be combined to obtain

y1 y2y2 − y1y2y2 − δy1y2y2 + λy2y2 − (δ + c)y1y2y2

+ (ρδ + ρ+ δ − δ2 − δc)y1y2y2 + cy2y2 + ρcy1y22 + (ρδc− δ2c)y1y2

2

−Nδy1y1y2 + cy1y22 −Nδ(ρ+ δ)y1y1y2

−Nδ2ρy12y2 +Nδ2λy1y2 = 0. (6.23)

The above equation just involves measurable state variables and constant parameters.Eq. (6.23) is of order 0 and of degree > 1 in θ, so Eq. (6.23) satisfies the third situationin Theorem 3.8 [64] in Section 3.3), thus θ = (λ, ρ,N, δ, c)T is locally identifiable.Consequently, η(t) is also locally identifiable. In addition, the identifiability of θ canalso be easily verified using the implicit function method based on Eq. (6.23). Theidentifiability of other similar HIV dynamic models have been studied in [56, 117, 119].

Page 29: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 29

6.3. Influenza A Virus Infection. The purpose of this section is to illustratepossible problems if identifiability analysis is ignored by considering influenza infectionin humans, an important infectious disease. Baccam et al. [8] proposed a target cell-limited model for influenza A virus infection:

dT

dt= −βTV, (6.24)

dI

dt= βTV − δI, (6.25)

dV

dt= pI − cV, (6.26)

where T is the number of uninfected target cells (epithelial cells), I is the number ofproductively infected cells, and V is the infectious viral titer expressed in TCID50/mlwhich is the only state variable to be measured. Since this is a low-dimension nonlineardynamic system, the implicit function method can be employed.

Considering the case that only V can be measured (and thus, for example, theinitial number of target cells T (0) is not known), we can derive the follow equationfrom Eq. (6.24)-(6.26) by eliminating the unmeasurable state variables:

V (3) =[V + δcV + (δ + c)V

](V −1V − βV )− δcV − (δ + c)V . (6.27)

Obviously, only the parameters (β, δ, c) can be identified and the minimum numberof required measurements of V is 6, and the parameter p is not identifiable in thiscase. Similarly, we consider the cases in which both I and V are measured, bothT and V are measured, or all three state variables are measured. For the casesthat any two or more state variables are measured, all four parameters (β, δ, c, p) arestructurally (theoretically) identifiable. We summarize the structural identifiabilityanalysis results for all cases in Table 6.3.

Baccam et al. [8] further proposed another target cell-limited influenza modelwith delayed virus production as follows:

dT

dt= −βTV, (6.28)

dI1dt

= βTV − kI1, (6.29)

dI2dt

= kI1 − δI2, (6.30)

dV

dt= pI2 − cV, (6.31)

where I1 is the number of latent infected epithelial cells that are not yet producingvirus and I2 the number of productively infected epithelial cells. Again, we canuse the implicit function theorem method to investigate the structural (theoretical)identifiability of this model. We summarize the identifiability analysis results in Table6.4. For this model, if only V is measured, four parameters (β, δ, c, k) are identifiableand parameter p is not identifiable. However, if any two or more of the four statevariables (T, I1, I2, V ) are measured, all five parameters (β, δ, c, k, p) are theoreticallyidentifiable. We also summarize the minimum number of required measurements foreach of the state variables in Table 6.4.

In the paper by Baccam et al. [8], only the virus titers were measured at 8 timepoints during days 1-8 of infection from six patients, but some of these measurements

Page 30: On Identifiability of Nonlinear Ode Models And

30 H. Miao, X. Xia, A.S. Perelson and H. Wu

Table 6.3Structural identifiability of the target cell-limited influenza model in Baccam et al. [8].

Measured Variables Identifiable Parameters Minimum number of measurementsV (β, δ, c) 6 of V

V and I (β, δ, c, p) 3 of V , 4 of IV and T (β, δ, c, p) 5 of V , 2 of T

V , I and T (β, δ, c, p) 3 of V , 2 of I, 2 of T

Table 6.4Structural identifiability of the target cell-limited influenza model with delayed virus production

in Baccam et al. [8].

Measured Variables Identifiable Parameters Minimum number of measurementsV (β, δ, c, k) 8 of V

V and T (β, δ, c, k, p) 7 of V , 2 of TV and I1 (β, δ, c, k, p) 5 of V , 4 of I1V and I2 (β, δ, c, k, p) 2 of V , 6 of I2

V , I1 and I2 (β, δ, c, k, p) 3 of V , I1 and I2V , T , I1 and I2 (β, δ, c, k, p) 3 of V , 2 of T , I1 and I2

were below detection. According to the identifiability analysis in Table 6.3 and Table6.4, p is not identifiable. To fit the four dimensional model (6.28)-(6.31), all 8 datapoints need to be used, otherwise more parameters may be unidentifiable. However,since the identification equation (6.27) does not involve the unidentifiable parameterp, one may fix p, which does not affect the estimates of other parameters. Sincethe identifiability analysis was not considered, the estimates of kinetic parameters inBaccam et al. [8] should be interpreted with caution since Baccam et al. [8] fixedT (0) to avoid the identifiability problem and T (0)’s value was not directly from data.

Influenza infection in human is a very complex problem, therefore, much morecomplicated models were proposed; however, the problem is that such work usuallyover-parameterizes the model and ignores parameter identifiability, which makes itdifficult to directly such models to data. For example, Hancioglu et al. [43] proposeda 10-equation model for influenza A virus infection (details not shown). We can showthat, to verify the identifiability of all the 27 parameters in that model, almost all the10 state variables need to be measured, which is nearly impossible to do in practicedue to technical and ethical limitation. In summary, when fitting model to data,models with parameter identifiability verified should be considered.

7. Discussion and Conclusion. Ordinary differential equations are an impor-tant tool for quantifying a dynamic process in many scientific fields, and recently ithas been widely used in modeling biomedical processes, in particular for modelinginfectious diseases and viral dynamics. It is critical to estimate the unknown kineticparameters in ODE models from experimental data in biomedical applications. How-ever, it is not trivial and not apparent to know whether the unknown parameters ingeneral nonlinear ODE models are identifiable based on the experimental data. Thus,an identifiability analysis is a prerequisite before any statistical method is applied toestimate the unknown parameters from the experimental data.

Three main categories of identifiability techniques have been developed for gener-al ODE models. The first is structural (theoretical) identifiability analysis, which canbe used to evaluate whether all parameters can be theoretically identified by manip-ulating the model structure. Two assumptions are needed for such analysis: 1) themodel structure is absolutely accurate; and 2) measurement is exact (no measurement

Page 31: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 31

error). Although these two assumptions are not realistic in practice, it is still neces-sary to study theoretical identifiability. The second type of identifiability analysis ispractical identifiability analysis in which both model uncertainty and practical mea-surement errors are considered. Structural identifiability analysis can be done beforeexperiments for data collection are designed. In fact, the structural identifiabilityanalysis can provide useful information, such as the minimum number of measure-ments at distinct time points, for experimental design. If an ODE model turns out tobe unidentifiable or only a subset of model parameters are identifiable via structuralidentifiability analysis, the model may need to be modified or some of the parametersmay need to be fixed before statistical methods are applied to estimate the unknownparameters. Otherwise, statistical estimates may not be reliable. Even though someparameter estimates can be obtained from an unidentifiable model, the estimates maybe a local estimates or an arbitrary set of estimates that can overfit the observationdata. If the structural identifiability analysis confirms that an ODE model is globallyor locally identifiable, practical identifiability analysis should be done to check thereliability and sensitivity of estimates to measurement errors and model uncertainty.Based on the results of practical identifiability analysis, a model can be further re-fined by model selection techniques [71]. Practical identifiability analysis can also beused to better design future experiments. The third type of identifiability analysistechnique is based on the sensitivity matrix. Similar to structural identifiability anal-ysis, sensitivity-based methods do not require experimental observations and cannotaccount for model uncertainty and measurement errors. Like practical identifiabilityanalysis, sensitivity-based methods also require at least one nominal value of param-eters. Note that so far it is still difficult to do structural identifiability analysis forhigh-dimensional ODEs or complicated ODEs. In this case, the practical identifiabil-ity analysis may not be reliable since the structural (theoretical) identifiability of themodel is unknown. The class of sensitivity-based methods, which is a technique be-tween the structural (theoretical) identifiability and practical identifiability analysis,can be used in such case.

Besides the identifiability analysis techniques for ODEs, the identifiability analysisof delay differential equation (DDE) models should also be discussed. A general formof a DDE system is given as follows

Mx(t) = f(t,x(t),x[τ (t,x)], x(t), x[τ (t,x)],u(t),θ), (7.1)

y(t) = h(x(t),x[τ (t,x)],u(t),θ), (7.2)

x(t0) = x(t0,u(t0),θ), (7.3)

where t0 is the starting value of the independent variable, x(t) ∈ Rm is a vector ofstate variables, y(t) ∈ Rd the measurable system output vector, u(t) ∈ Rp the knownsystem input vector, θ ∈ Rq the parameter vector, M a constant coefficient matrix(or mass matrix), and τ (t,x) a vector of delay functions. It is required that

τ (t,x) ≤ t, (7.4)

that is, the value of delay functions should always be smaller than or equal to thecurrent time which is reasonable since the future value is unknown yet. Anotherimportant assumption is that

rank(∂h

∂x) = d, (7.5)

Page 32: On Identifiability of Nonlinear Ode Models And

32 H. Miao, X. Xia, A.S. Perelson and H. Wu

which implies that none of the d system outputs is trivial (e.g., linear combinationof other outputs). Since τ (t,x) ≤ t, it is necessary to know the value of x(t) whent ≤ t0, i.e., the history function:

x(t) = g(t,θ), if t ≤ t0 (7.6)

where x(t), t ≤ t0 is only a function of time and parameters. It is obvious thatthe DDE system described here is much more complicated than an ODE system,and currently it is still impossible to numerically solve a very general DDE model.However, for some relatively simple DDE models, a number of numerical methods havebeen proposed and implemented. For details of such algorithms, the reader is referredto the work by Ascher and Petzold [6], Bellen and Zennaro [11], Guglielmi and Hairer[41, 42], and Shampine and Thompson [97]. Particularly, Guglielmi and Hairer [41, 42]proposed and implemented a comparatively general solver (called Radau IIA) forDDE models, which is recommended for practical applications due to its efficiency andstability. For examples of DDE modeling of HIV infection, the reader is referred to [90]and [76]. A number of independent studies have tackled the identifiability problem forDDE models [4, 5, 27, 31, 66, 68, 79, 124, 125]. However, most of these previous worksdeal with very simple and specific DDE models (e.g., [68]) and the generalizabilityof these results are limited due to a lack of understanding of the important featureof DDE models: the propagation of discontinuity at t0 or in the history functionsfrom lower to higher order derivatives of state variables if there is any. Such a featuremakes DDE models easily become bifurcated or just unsolvable [11, 41, 42]. Theidentifiability conclusions based on model structure manipulation are not reliableunless analytical solutions of DDE models can be obtained and analyzed. However,it is surprising that almost all the existing work tried to tackle the identifiabilityproblem by manipulating model structures (e.g., [5]). Generally, for complicatedsystems as described in Eqs. (7.1) to (7.3), it is extremely difficult to manipulatethe model structures to study identifiability problems. Thus, the methodologies forDDE model identifiability are still in its infancy and promising approaches are likelyto be numerical methods such as the practical or sensitivity-based methods (e.g.,[9]), although it may require development and reliable realization of DDE numericalalgorithms.

In addition, it should be mentioned that some of the identifiability techniquessuch as the differential algebra method can be extended to study the identifiability ofpartial differential equation (PDE) models (e.g., [49]), but the identifiability analysisfor more complicated models such as PDEs or stochastic differential equations isbeyond the scope of this paper.

Finally, after the identifiability analysis is done, statistical estimation methodsshould be used to estimate the unknown parameters in the model. The practicalidentifiability analysis also requires reliable parameter estimation methods be avail-able. Recently statistical estimation methods for ODE models have attracted a greatdeal of attention from statisticians. Some novel and efficient estimation methods par-ticular for nonlinear ODE models have been published in the statistical literature[17, 18, 40, 48, 61, 62, 88]. Besides the standard least squares approach [70, 71],more reliable and computationally efficient estimation methods and their theoreticalfoundations have been developed [17, 18, 62, 88]. However, the topic of statisticalestimation methods for ODE, DDE and PDE models is beyond the scope of thispaper.

Page 33: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 33

Acknowledgments. This work was partially supported by NIAID/NIH researchgrants AI055290, AI50020, AI28433, AI078498, RR06555, University of RochesterProvost Award, and University of Rochester DCFAR (P30AI078498) Mentoring Award.

Page 34: On Identifiability of Nonlinear Ode Models And

34 H. Miao, X. Xia, A.S. Perelson and H. Wu

REFERENCES

[1] B. M. Adams, H. T. Banks, H. D. Kwon, and H. T. Tran, Dynamic multidrug therapies forHIV: Optimal and STI control approaches, Mathe. Biosci. Eng., 1 (2004), pp. 223–241.

[2] S. Aluru, Haendbook of Computational Molecular Biology (Chapman & All/Crc Computerand Information Science Series), Chapman & Hall/CRC, Boca Raton, FL, 2005.

[3] D. H. Anderson, Compartmental modeling and tracer kinetics, S. Levin, eds., Lecture Notesin Biomathematics, Vol.50, Springer, Berlin, 1983.

[4] M. Anguelova, Observability and identifiability of nonlinear systems with applications in bi-ology, PH.D. thesis, Department of Mathematical Sciences, Chalmers University of Tech-nology and Goteborg University, Goteborg, Sweden, 2007.

[5] M. Anguelova and B. Wennberg, State elimination and identifiability of the delay param-eter for nonlinear time-delay systems, Automatica, 44 (2008), pp. 1373–1378.

[6] U. M. Ascher and L. R. Petzold, The numerical solution of delay-differential-algebraicequations of retarded and neutral type, SIAM J. Numer. Anal., 32 (1995), pp. 1635–1657.

[7] S. Audoly, G. Bellu, L. D’Angio, M. P. Saccomani, and C. Cobelli, Global identifiabilityof nonlinear models of biological systems, IEEE Trans. Biomed. Eng., 48 (2001), pp. 55–65.

[8] P. Baccam, C. Beauchemin, C. A. Macken, F. G., Hayden, A. S. Perelson, Kinetics ofinfluenza A virus infection in humans, J. Virol., 80 (2006), pp. 7590–7599.

[9] H. T. Banks and D. M. Bortz, A parameter sensitivity methodology in the context of HIVdelay equation models, J. Math. Biol., 50 (2005), pp. 607–625.

[10] R. Bellman and K. J. Astrom, On structural Identifiability, Math. Biosci., 7 (1970), pp. 329–339.

[11] A. Bellen and M. Zennaro, Numerical solution of delay differential equations, OxfordUniversity Press, Oxford, 2003.

[12] G. Bellu, M. P. Saccomani, S. Audoly, and L. D’Angi, DAISY: A new software tool to testglobal identifiability of biological and physiological systems, Comput. Methods ProgramsBiomed., 88 (2007), pp. 52–61.

[13] D. G. Cacuci, Sensitivity and Uncertainty Analysis: Theory. Volume 1, Chapman Hall/CRC,Boca Raton, FL, 2003.

[14] G. Carra’Ferro and V. P. Gerdt, Improved Kolchin-Ritt algorithm, Program. Comput.Soft. 29 (2003), pp. 83–87.

[15] C. Castillo-Chavez, S. Blower, P. vande Driessche, D. Kirschner, and A. A. Yakubu,Mathematical approaches for emerging and reemerging infectious diseases: models, meth-ods, and theory, Springer, New York, 2002.

[16] M. J. Chappel and K. R. Godfrey, Structural identifiability of the parameters of a nonlinearbatch reactor model, Math. Biosci., 108 (1992), pp. 245–251.

[17] J. Chen and H. Wu, Estimation of time-varying parameters in deterministic dynamic modelswith application to HIV infections, Statistica Sinica, 18 (2008a), pp. 987–1006.

[18] J. Chen and H. Wu, Efficient local estimation for time-varying coefficients in deterministicdynamic models with applications to HIV-1 dynamics, J. Amer. Stat. Assoc., 103 (2008b),pp. 369–384.

[19] C. Cobelli and G. Romanin-Jacur, Controllability, observability and structural identifi-ability of multi input and multi output biological compartmental systems, IEEE Trans.Biomed. Eng., 23 (1976), pp. 93–100.

[20] C. Cobelli, A. Polo, and G. Romanin-Jacur, A computer program for the analysis ofcontrollability, observability and structural identifiability of biological compartmental sys-tems, Comput. Programs Biomed., 7 (1977), pp. 21–36.

[21] C. Cobelli, A. Lepschy, and R. Jacur, Identifiability of compartmental systems and relatedstructural proerties, Math. Biosci., 44 (1979), pp. 1–18.

[22] J. M. Coffin, HIV Population Dynamics in Vivo: Implications for Genetic Variation, Patho-genesis, and Therapy, Science, 267 (1995), pp. 483–489.

[23] E. J. Davidson, Connectability and structural controllability of composite systems, Automat-ica, 13 (1997), pp. 109–123.

[24] D. Degenring, C. Froemel, G. Dikta, and R. Takors, Sensitivity analysis for the reductionof complex metabolism models, J. Proc. Cont., 14 (2004), pp. 729–745.

[25] L. Denis-Vidal and G. Joly-Blanchard, An easy to check criterion for (un)identifiability ofuncontrolled systems and its applications, IEEE Trans. Auto. Cont., 45 (2000), pp. 768–771.

[26] L. Denis-Vidal, G. Joly-Blanchard, and C. Noiret, Some effective approaches to checkthe identifiabiity of uncontrolled nonlinear systems, Math. Comput. Simul., 57 (2001),

Page 35: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 35

pp. 35–44.[27] L. Denis-Vidal, C. Jauberthie, and G. Joly-Blanchard, Identifiability of a nonlinear

delayed-differential aerospace model, IEEE Trans. Auto. Cont., 51 (2006), pp. 154–158.[28] W. E. Deskins, Abstract algebra, Dover, 1996.[29] S. Diop and M. Fliess, On nonlinear observability, in Proc. of the first Europ. Cont. Conf.,

Paris, Hermes, 1991, pp. 152–157.[30] E. P. Dougherty, J. T. Hwang, and H. Rabitz, Further developments and applications of

the Green’s function method of sensitivity analysis in chemical kinetics, J. Chem. Phy.,71 (1979), pp. 1794–1808.

[31] G. Ferretti, C. Maffezzoni, and R. Scattolini, On the identifiability of the time delaywith least squares methods, Automatica, 32 (1996), pp. 449–453.

[32] R. A. Filter, X. Xia, and C. M. Gray, Dynamic HIV/AIDS parameter estimation withapplication to a vaccine readiness study in Southern Africa, IEEE Trans. Biomed. Eng.,52 (2005), pp. 784–791.

[33] B. R. Frieden, Science from Fisher Information, Cambridge University Press, New York,2004.

[34] C. Froemel, Parameterreduktion in stoffwechselmodellen mit methoden der statistik, Masterthesis, Fachhochschule Aachen, Abteilung Juelich, Fachbereich Physikalische Technik,Studienrichtung Technomathematik, 2003.

[35] K. G. Gadkar, R. Gunawan, F. J. Doyle, Iterative approach to model identification ofbiological networks, BMC Bioinformatics, 6 (2005), pp. 155–175.

[36] R. Gilmore, Lie groups, Lie algebras, and some of their applications, Dover, 2006.[37] S. T. Glad, Solvability of differential algebraic equations and inequalities: an algorithm, in

Europ. Cont. Conf., Brussels, 1997.[38] W. J. Grantham and T. L. Vincent, Modern control systems analysis and design, Wiley,

New York, 1993.[39] M. Grewal and K. Glover, Identifiability of linear and nonlinear dynamical systems, IEEE

Trans. Autom. Cont., 21 (1976), pp. 833–837.[40] J. Guedj, R. Thiebaut, and D. Commenges, Practical identifiability of HIV dynamics mod-

els, Bull. Math. Biol., 69 (2007), pp. 2493–2513.[41] N. Guglielmi and E. Hairer, Implementing Radau IIA methods for stiff delay differential

equations, Computing, 67 (2001), pp. 1–12.[42] N. Guglielmi and E. Hairer, Computing breaking points in implicit delay differential equa-

tions, Adv. Comput. Math., 29 (2008), pp. 229–247.[43] B. Hancioglu, D. Swigon, and G. Clermont, A dynamical model of human immune re-

sponse to influenza A virus infection, J. Theor. Biol., 246 (2007), pp. 70–86.[44] R. Hermann and A. J. Krener, Nonlinear controllability and observability, IEEE Trans.

Autom. Cont., 22 (1977), pp. 728–740.[45] D. D. Ho, A. U. Neumann, A. S. Perelson, W. Chen, J. M. Leonard, and M. Markowitz,

Rapid turnover of plasma virions and CD4 lymphocytes in HIV-1 infection, Nature, 373(1995), pp. 123–126.

[46] A. Holmberg, On the practical identifiability of microbial growth models incorporatingMichaelis-Menten type nonlinearities, Math. Biosci., 62 (1982), pp. 23–43.

[47] K. Hsu, C. Novara, and T. Vincent, M. Milanese, and K. Poolla, Parametric andnonparametric curve fitting, Automatica, 42 (2006), pp. 1869–1873.

[48] Y. Huang, D. Liu, and H. Wu, Hierarchical Bayesian methods for estimation of parametersin a longitudinal HIV dynamic system, Biometrics, 62 (2006), pp. 413–423.

[49] E. Hubert, Essential components of algebraic differential equations, J. Symb. Comput., 28(1999), pp. 657–680.

[50] A. Isidori, Nonlinear control systems, 3rd ed., Springer, New York, 1995.[51] J. A. Jacquez, Compartmental analysis in biology and medicine, 2nd ed., The University of

Michigan Press, Ann Arbor, Michigan, 1985.[52] J. A. Jacquez and P. Greif, Numerical parameter identifiability and estimability: Integrat-

ing identifiability, estimability, and optimal sampling design, Math. Biosci., 77 (1985),pp. 201–227.

[53] L. Jaulin, M. Kieffer, O. Didrit, and E. Walter, Applied Interval Analysis, Springer-Verlag, London, 2001.

[54] I. T. Jolliffe, Discarding variables in a principal component analyssis. I: artificial data, J.Royal Stat. Soc., 21 (1972), pp. 160–172.

[55] A. M. Jeffrey, X. Xia, and I. K. Craig, When to initiate HIV therapy: a control theoreticapproach, IEEE Trans. Biomed. Eng., 50 (2003), pp. 1213–1220.

[56] A. M. Jeffrey and X. Xia, Identifiability of HIV/AIDS model, in Deterministic and stochas-

Page 36: On Identifiability of Nonlinear Ode Models And

36 H. Miao, X. Xia, A.S. Perelson and H. Wu

tic models of AIDS epidemics and HIV infections with intervention, W. Y. Tan and H.Wu, eds., World Scientific, 2005.

[57] E. Kolchin, Differential algebra and algebraic groups, Academic Press, New York, 1973.[58] O. Krakovska and L. M. Wahl, Optimal drug treatment regimens for HIV depend on

adherence, J. Theoret. Biol., 246 (2007), pp. 499–509.[59] D. A. Lauffenburger and J. L. Linderman, Receptors: models for binding, trafficking, and

signaling, Oxford University Press, Oxford, 1993.[60] W. S. Levine, The control handbook, CRC Press, 1996.[61] L. Li, M. B. Brown, K. H. Lee, and S. Gupta, Estimation and inference for s spline-

enhanced population pharmacokinetic model, Biometrics, 58 (2002), pp. 601–611.[62] H. Liang and H. Wu, Parameter estimation for differential equation models using a frame-

work of measurement error in regression model, J. Amer. Stat. Assoc., 103 (2008), p-p. 1570–1583.

[63] L. Ljung, Convergence analysis of parametric identification methods, IEEE Trans. Autom.Cont., 23 (1978), pp. 770–783.

[64] L. Ljung and T. Glad, On global identifiability for arbitrary model parametrizations, Auto-matica, 30 (1994), pp. 265–276.

[65] L. Ljung, System identification: theory for the user, 2nd ed., Prentice Hall, New Jersey, 1999.[66] S. M. V. Lunel, Parameter identifiability of differential delay equations, Int. J. Adapt. Cont.

Sig. Proc., 15 (2001), pp. 655–678.[67] E. L. Mansfield and P. A. Clarkson, Applications of the differential algebra package d-

iffgrob2 to classical symmetries of differential equations, J. Symb. Comput., 23 (1997),pp. 517–533.

[68] J. A. Merino, J. D. Biasi, Y. Plusquellec, and G. Houin, Local identifiability for two andthree-compartment pahrmacokinetic models with time-lags, Med. Eng. Phy., 20 (1998),pp. 261–268.

[69] N., Metropolis and S. Ulam, The Monte Carlo Method, J. Amer. Stat. Assoc., 44 (1949),pp. 335–341.

[70] H. Miao, C. Dykes, L. M. Demeter, A. S. Perelson, and H. Wu, Modeling and estimationof kinetic parameters and replication fitness of HIV-1 from Flow-Cytometry-based growthcompetition experiments, Bull. Math. Biol., 70 (2008), pp. 1749–1771.

[71] H. Miao, C. Dykes, L. M. Demeter, and H. Wu, Differential equation modeling of HIV viralfitness experiments: Model identification, model Selection, and multi-model inference,Biometrics, 65 (2009), pp. 292–300.

[72] B. Mishra, Algorithmic algebra, Springer, New York, 1993.[73] J. R. Munkres, Analysis on Manifolds, Addison-Wesley, 1991.[74] J. D. Murray, Mathematical biology I: An introduction, Springer, New York, 2002.[75] J. D., Murray, Mathematical biology II: Spatial models and biomedical applications, Springer,

New York, 2003.[76] P. W. Nelson and A. S. Perelson, Mathematical analysis of delay differential equation

models of HIV-1 infection, Math. Biosci., 179 (2002), pp. 73–94.[77] M. A. Nowak and R. M. May, Virus dynamics: mathematical principles of immunology and

virology, Oxford University Press, Oxford, 2000.[78] F. Ollivier, Le probleme de l’identifiabilite globale: etude the orique, methodes effectives et

bornes de complexite, Ph.D. Thesis, ecole Polytechnique, Paris, France, 1990.[79] Y. Orlov, L. Belkoura, J. P. Richar, and M. Dambrine, On identifiability of linear

time-delay systems, IEEE Trans. Auto. Cont., 47 (2002), pp. 1319– 1324.[80] A. S. Perelson, P. Essunger, and Y. Cao et al., Decay characteristics of HIV-1-Infected

compartments during combination therapy, Nature, 387 (1997), pp. 188–191.[81] A. S. Perelson, Modelling viral and immune system dynamics, Nat. Rev. Immunol., 2 (2002),

pp. 28–36.[82] A. S. Perelson and P. W. Nelson, Mathematical analysis of HIV-1 dynamics in vivo, SIAM

Rev., 41 (1999), pp. 3–44.[83] M. Pia Saccomani, S. Audoly, and L. D’Angi, Parameter identifiability of nonlinear sys-

tems: the role of initial conditions, Automatica, 39 (2003), pp. 619–632.[84] H. Pohjanpalo, System identifiability based on the power series expansion of the solution,

Math. Biosci., 41 (1978), pp. 21–33.[85] T. Quaiser, W. Marquardt, and M. Monnigmann, Local identifiability analysis of large

signaling pathway models, in 2nd Conference Foundation of Systems Biology in Engineer-ing, Proceedings Plenary and Contributed Papers, ISBN 978-3-8167-7436-5, F. Allgoewerand M. Reuss, eds., Fraunhofer IRB Verlag, Stuttgart, Germany, 2007, pp. 465–470.

[86] T. Quaiser and M. Monnigmann, Systematic identifiability testing for unambiguous mecha-

Page 37: On Identifiability of Nonlinear Ode Models And

Identifiability of ODE Models 37

nistic modeling - application to JAK-STAT, MAP kinase, and NF-κB signaling pathwaymodels, BMC Syst. Biol., 3 (2009), pp. 50–71.

[87] A. Raksanyi, Y. Lecourtier, E. Walter, and E. Venot, Identifiability and distinguisha-bility testing in computer algebra, Math. Biosci., 77 (1985), pp. 245–266.

[88] J. O. Ramsay, G. Hooker, D. Campbell, and J. Cao, Parameter estimation for differentialequations: A generalized smoothing approach (with discussions), J. Roy. Stat. Soc. B, 69(2007), pp. 741–796.

[89] C. R. Rao, Information and the accuracy attainable in the estimation of statistical parame-ters, Bull. Calcutta Math. Soc., 37 (1945), pp. 81–91.

[90] V. C. Rebecca and S. Ruan, A delay-differential equation model of HIV infection of CD4+T-cells, Math. Biosci., 165 (2000), pp. 27–39.

[91] J. F. Ritt, Differential Algebra, American Mathematical Society, Providence, RI, 1950.[92] M. Rodriguez-Fernandez, J. A. Egea, and J. R. Banga, Novel metaheuristic for parame-

ter estimation in nonlinear dynamic biological systems, BMC Bioinformatics, 7 (2006a),pp. 483–500.

[93] M. Rodriguez-Fernandez, P. Mendes, and J. R. Banga, A hybrid approach for efficientand robust parameter estimation in biochemical pathways, Biosystems, 83 (2006b), p-p. 248–265.

[94] A. Saltelli, K. Chan, and M. Scott, Sensitivity analysis, Wiley Series in Probability andStatistics, John Wiley & Sons, New York, 2000.

[95] C. Seigneur, G. Stephanopoulos, and R. W. Carr, Dynamic sensitivity analysis of chem-ical reaction system: a variational method, Chem. Eng. Sci., 37 (1982), pp. 845–853.

[96] K. Schittkowski, Experimental design tools for ordinary and algebraic differential equations,Ind. Eng. Chem. Res., 46 (2007), pp. 9137–9147.

[97] L. F. Shampine and S. Thompson, Solving DDEs in MATLAB, App. Num. Math., 37 (2001),pp. 441–458.

[98] E. D. Sontag, On the observability of polynomial systems, SIAM J. Cont. Opt., 17 (1979),pp. 139–151.

[99] E. D. Sontag, Spaces of observables in nonlinear control, in Proc. Int. Congress of Mathe-maticians, Birkhauser Verlag, Basel, 2 (1995), pp. 1532–1545.

[100] E. D. Sontag, For differential equations with r parameters, 2r + 1 experiments are enoughfor identification, J. Nonlin. Sci., 12 (2002), pp. 553–583.

[101] E. D. Sontag, Y. Wang, and A. Megretski, Input classes for identification of bilinearsystems, IEEE Trans. Autom. Cont., 54 (2009), pp. 195–207.

[102] F. Takens, Detecting strange attractors in turbulence, in Proc. Symp. Dyna. Syst. Turbulence,D. A. Rand and L. S. Young, eds., Springer-Verlag, Berlin, 1981.

[103] K. Thomaseth and C. Cobelli, Generalized sensitivity functions in physiological systemidentification, Ann. Biomed. Eng., 27 (1999), pp. 607–616.

[104] T. Tunali and T. J. Tarn, New results for identifiability of nonlinear systems, IEEE Trans.Autom. Contorl., 32 (1987), pp. 146–154.

[105] T. Turanyi, Sensitivity analysis of complex kinetic systems. Tools and applications, J. Math.Chem., 5 (1990), pp. 203–248.

[106] S. Vajda, K. Godfrey, and H. Rabitz, Similarity transformation approach to identifiabilityanalysis of nonlinear compartmental models, Math. Biosci., 93 (1989a), pp. 217–248.

[107] S. Vajda and H. Rabitz, State isomorphism approach to global identifiability of nonlinearsystems, IEEE Trans. Autom. Cont., 34 (1989), pp. 220–223.

[108] S. Vajda, H. Rabitz, E. Walter, and Y. Lecourtier, Qualitative and quantitative identifi-ability analysis of nonlinear chemical kinetiv-models, Chem. Eng. Commun., 83 (1989b),pp. 191–219.

[109] H. L. Van Trees, Detection, estimation, and modulation theory, Part I, Wiley, New York,1968.

[110] R. Vidal, A. Chiuso, and S. Soatto, Observability and identification of jump linear systems,in Proceedings of 41st IEEE conference on decision and control, 41 (2002), pp. 3614–3619.

[111] T. L. Vincent and W. J. Grantham, Nonlinear and optimal control systems, Wiley-Interscience, New York, 1997.

[112] E. Walter, I. Braems, L. Jaulin, and M. Kieffer, Guaranteed numerical computation asan alternative to computer algebra for testing models for identifiability, Lecture Notes inComputer Science, 2004, pp. 124–131.

[113] E. Walter and Y. Lecourtier, Unidentifiable compartmental models: what to do?, Math.Biosci., 56 (1981), pp. 1–25.

[114] E. Walter, Identifiability of parameteric models, Pergamon Press, Oxford, 1987.[115] H. Whitney, Differentiable manifolds, Ann. Math., 37 (1936), pp. 645–680.

Page 38: On Identifiability of Nonlinear Ode Models And

38 H. Miao, X. Xia, A.S. Perelson and H. Wu

[116] H. Whitney, The self-intersections of a smooth n-manifold in 2n-space, Ann. Math., 45(1944), pp. 220–246.

[117] H. Wu, H. Zhu, H. Miao, and A. S. Perelson, Identifiability and statistical estimation ofdynamic parameters in HIV/AIDS dynamic models, B. Math. Biol., 70 (2008), pp. 785–799.

[118] X. Xia, Estimation of HIV/AIDS parameters, Automatica, 39 (2003), pp. 1983–1988.[119] X. Xia and C. H. Moog, Identifiability of nonlinear systems with applications to HIV/AIDS

models, IEEE Trans. Autom. Contorl., 48 (2003), pp. 330–336.[120] K. Z. Yao, B. M. Shaw, B. Kou, K. B. McAuley, and D. W. Bacon, Modeling ethy-

lene/butene copolymerization with multi-site catalysts: Parameter estimability and ex-perimental design, Polym. React. Eng., 11 (2003), pp. 563–588.

[121] H. Yue, M. Brown, J. Knowles, H. Wang, D. S. Broomhead, and D. B. Kell, Insightsinto the behaviour of systems biology models from dynamic sensitivity and identifiablityanalysis: a case study of an NF-kappaB signalling pathway, Mol. BioSyst., 2 (2006),pp. 640–649.

[122] D. E. Zak, G. E. Gonye, J. S. Schwaber, and F. J. Doyle, Importance of input pertur-bations and stochastic gene expression in the reverse engineering of genetic regulatorynetworks: insights from an identifiability analysis of an in silico network, Genome Res.,13 (2003), pp. 2396–2405.

[123] R. M. Zazworsky and H. K. Knudsen, Comments on “Controllability, observability andstructural identifiability of multi input and multi output biological compartmental systems,IEEE Trans. Biomed. Eng., 24 (1977), pp. 495–496.

[124] J. Zhang, X. Xia, C. H. Moog, Parameter identifiability of nonlinear systems with time-delay, IEEE Trans. Auto. Cont., 51 (2006), pp. 371– 375.

[125] J. Zhang and X. Xia, Identifiability problems of time-delay HIV models, in 17th IFAC WorldCongress, Seoul, Korea, 2008.

[126] R. Zurakowski and A. R. Teel, A model predictive control based scheduling method forHIV therapy, J. Theoret. Biol., 238 (2006), pp. 368–382.