HomeJournalsSearchAlertsAccountHelp
2 of 9 Article List Previous Next


Automatica
Volume 36, Issue 9 SummaryPlus
September 2000 Article
Pages 1249-1274 Journal Format-PDF (293 K)


PII: S0005-1098(00)00050-9
Copyright © 2000 Elsevier Science B.V. All rights reserved.

Survey Paper

A survey of computational complexity results in systems and control*1

Vincent D. Blondel, , a and John N. Tsitsiklis, b

a Department of Mathematical Engineering, Center for Systems Engineering and Applied Mechanics (CESAME), University of Louvain, B-1348 Louvain-la-Neuve, Belgium
b Laboratory for Information and Decision Systems, Massachusetts Institute of Technology, Cambridge, MA 02139, USA

Received 4 November 1998; revised 16 July 1999 and 6 October 1999. Available online 25 May 2000.

Abstract

The purpose of this paper is twofold: (a) to provide a tutorial introduction to some key concepts from the theory of computational complexity, highlighting their relevance to systems and control theory, and (b) to survey the relatively recent research activity lying at the interface between these fields. We begin with a brief introduction to models of computation, the concepts of undecidability, polynomial-time algorithms, NP-completeness, and the implications of intractability results. We then survey a number of problems that arise in systems and control theory, some of them classical, some of them related to current research. We discuss them from the point of view of computational complexity and also point out many open problems. In particular, we consider problems related to stability or stabilizability of linear systems with parametric uncertainty, robust control, time-varying linear systems, nonlinear and hybrid systems, and stochastic optimal control.

Author Keywords: Control; Discrete-event systems; Discrete-time systems; Hybrid systems; Markov decision processes; Mathematical systems theory; Neural networks; Nonlinear systems; Time-varying systems; Turing machines

Article Outline

1. Introduction
2. Complexity theory
2.1. Problems, instances, and the size of an instance
2.2. Algorithms
2.3. Decidable and undecidable problems
2.4. Time complexity
2.5. Turing machines and other models
2.6. Some polynomial-time solvable problems
2.7. NP-completeness and beyond
2.8. Coping with NP-completeness
3. Linear systems
3.1. Stable and unstable polynomials in polynomial families
3.2. Stability of matrix families
3.3. Structured singular values
3.4. Existence of a stable matrix in an interval family; static output feedback
3.5. Time-varying systems
4. Nonlinear and hybrid systems
4.1. Introduction
4.2. Systems with a single nonlinearity
4.3. Systems with componentwise nonlinearities
4.4. Piecewise linear and hybrid systems
4.5. Other classes of systems
5. Stochastic control
5.1. Perfectly observed MDPs
5.2. Succintly described problems
5.3. Problems with continuous state spaces
5.4. Problems with imperfect information
5.5. Nonclassical information structures
5.6. Supervisory control of discrete-event systems
5.7. Zero-sum Markov games
6. Conclusions
Acknowledgements
References
Vitae

1. Introduction

In order to motivate the reader, and as a preview of the types of problems and "solutions" overviewed in this paper, we start with a deceptively simple problem.

We are given two n×n real matrices A+ and A-, and we consider products of the form

AiT···Ai2Ai1,(1)

where T is an arbitrary positive integer and (i1,i2,...,iT) is an arbitrary finite sequence whose elements take values in the set {+,-}. We wish to decide whether there exists some T and a sequence of length T such that the corresponding product is a stable matrix (in the sense that all of its eigenvalues have magnitude strictly less than 1).

If we are given an upper bound T* on the allowed values of T, the problem has an obvious solution: form all the possible products and check whether one of them is stable. The amount of computation required by this solution grows exponentially with T*. We will see in Section 3.5 that the exponential growth of the computational resources needed for a solution is probably unavoidable.

In the absence of a bound T*, we have to form and check an infinite set of matrix products, which cannot be done in finite time under any reasonable model of computation. We may wonder whether a shortcut is possible that only involves a finite amount of computation. It turns out that none exists, and the problem is unsolvable in a precise mathematical sense; see Section 3.5.

By turning around the question of the preceding paragraph, we could ask whether all products (1) are stable. As will be discussed in Section 3.5, this is related to some major open problems: an algorithmic solution is not known, but a proof of unsolvability is not available either.

In this paper, we introduce background material that will make statements such as the above precise, and also provide a review of existing results of this type. The key concepts that we will be working with refer to the algorithmic solvability of different problems and to the amount of resources (e.g., computation time) required. The premise behind our development is that a problem has been "solved" only if an algorithm for that problem is available. This viewpoint is an outgrowth of the pioneering work of Church, Gödel, and Turing earlier in this century, and has been refined through an additional focus on computation time requirements. It meshes very well with the availability of powerful computing machines, which has brought sophisticated systems and control approaches into the realm of the practical.

The remainder of the paper is organized as follows. In Section 2, we present models of digital computation and develop the notion of undecidability. We also introduce the concepts of polynomial-time algorithms and NP-completeness. We present some basic problems that are known to be undecidable or NP-complete, and conclude by discussing the consequences of NP-completeness results from a pragmatic point of view. In Section 3, we survey complexity-theoretic results that relate to linear systems, with a focus on time-invariant or time-varying uncertainty. In Section 4, we consider several classes of nonlinear systems, such as systems with a single nonlinearity, linear systems with componentwise nonlinearities, and hybrid systems. In Section 5, we move to stochastic optimal control problems and Markov decision theory, together with some pointers to related literature on discrete-event systems. We end with some brief conclusions in Section 6. Throughout the paper, several open problems are pointed out.

The literature on these subjects is large and growing, and our presentation could not be exhaustive. Additional information can be found in a number of more focused reviews such as Sontag (1995) (nonlinear control), Blondel and Tsitsiklis (1998) (elementary nonlinear systems), Littman, Dean and Kaelbling (1995), Rust (1996), Mundhenk, Goldsmith, Lusena and Allender (1997), (Markov decision theory), Alur et al. (1995) (hybrid systems), Bournez and Cosnard (1996), Moore (1998) (Turing machine as dynamical systems) and Orponen (1994) (recurrent neural networks). Some open problems related to the computational complexity of control questions are proposed in Blondel, Sontag, Vidyasagar and Willems (1999c, Problems 11, 14, 22, and 43).

2. Complexity theory

In this section, we introduce the main concepts from the theory of computability and complexity, with a focus on models of digital computation. An accessible introduction to the concepts of decidability/undecidability is given in Jones (1974). For textbook treatments, we refer the reader to Minsky (1967), Aho, Hopcroft and Ullman (1974), Garey and Johnson (1979), Hopcroft and Ullman (1979), Davis (1982) and Papadimitriou (1994). There is also an extensive literature on algebraic and continuous models of computation, and on problems with continuous data, which we do not cover; we refer the reader to Nemirovsky and Yudin (1983), Traub, Wasilkowski and Wozniakowski (1988) and Blum, Cucker, Shub and Smale (1998).

2.1. Problems, instances, and the size of an instance

In a typical instance of a computational problem, we are given input data x1,...,xd and we are asked to compute some function of them. For example, in the problem, an instance consists of the entries of a given matrix and the desired outcome is its determinant. We focus on digital computation and constrain the input data to be given with a finite number of bits. For example, we will typically assume that the inputs take integer or rational values. As far as negative complexity results are concerned, this is hardly restrictive. If a function is difficult or impossible to evaluate on integer or rational inputs, then it certainly remains so with arbitrary real inputs.

Different instances of the same problem can have different "sizes". We follow the standard convention of mainstream complexity theory and define the size of an instance as the number of bits used to encode the input data according to a certain prespecified format. In particular, any (nonzero) integer i can be viewed as having size (bit length) approximately equal to log |i|, since this is roughly the number of bits in a binary representation of i. For example, the size of an instance of the problem involving an n×n integer matrix can be upper bounded by O(n2 log U), where U is an upper bound on the magnitude of the entries of the matrix.1

2.2. Algorithms

Loosely speaking, an algorithm is described by a program, i.e., a finite sequence of instructions of the type encountered in common programming languages. For a more precise description, we need to specify a model of computation. A rather simple such model is provided by a random access machine (RAM) (Aho et al., 1974; Papadimitriou, 1994). A RAM consists of a read-only tape that contains the input data x1,...,xd, an output tape on which the outcome of the computation is written, an unlimited sequence of registers used to store intermediate quantities generated in the course of the computation, and a program. Each register and each memory location on the input and output tapes may contain an arbitrary, possibly negative, integer. The program is a sequence of instructions, some of which may be associated with labels that are used in "jump to" instructions. The instruction set can be assumed to contain the halting instruction, the four arithmetic operations, instructions for reading from a location on the input tape into a register (respectively, writing the contents of a register to a location on the output tape), indirect addressing (e.g., read the contents of the register whose address is stored in register i, and store them in register 1), and jump instructions that move the program to a next instruction, possibly depending on the outcome of the comparison of some register to zero. The exact nature of the allowed instructions is not important, because alternative choices of the instruction set lead to computing machines with equivalent computing capabilities, as long as all critical abilities, such as jumping, are present.

In a typical execution of the algorithm (or "computation"), the input data are loaded on the input tape, and the algorithm starts carrying out the program instructions. For any given input, the computation may or may not halt (that is, reach a halting instruction). We say that the algorithm solves a particular problem if it always halts (for every instance of the problem, that is, for every choice of the input data) and produces the correct answer on the output tape. We also say that a problem is unsolvable if there exists no algorithm (under our model of computation) that will always halt with the correct answer.

One may wonder whether the class of solvable problems depends on the choice of a model of computation. According to the so-called Church¯Turing thesis, all reasonable models of digital computation lead to the same class of solvable problems, and are therefore equally powerful. This thesis is supported by the fact that all reasonable models that have been proposed and studied lead indeed to the same class of solvable problems. Still, it only remains a thesis ¯¯ rather than a theorem ¯¯ because we do not have a precise definition of "all reasonable models".

2.3. Decidable and undecidable problems

We will now focus on so-called decision problems, that is, problems where the desired output is binary, and can be interpreted as "yes" or "no". An example is the problem of deciding whether a given (integer) matrix is nonsingular, which can be solved by, say, computing its determinant and comparing it to zero. This makes a decidable problem, one for which there exists an algorithm that always halts with the right answer. Similar to , many problems in linear algebra are decidable. But there also exist undecidable problems, for which there is no algorithm that always halts with the right answer. We give below one additional example of a decidable problem and then describe three examples of undecidable problems.

Let us consider the following problem. We are given a set of multivariable polynomials p1,...,pm in n real variables, with rational coefficients. Let S be the set of all x=(x1,...,xn)Rn that satisfy equalities and inequalities of the form pi(x1,...,xn)=0, pj(x1,...,xn)>0, pk(x1,...,xn)0. The problem of deciding whether S is nonempty (i.e., whether there exists a solution) is known to be decidable. More generally, we may consider questions such as whether the following statement is true:

x1,x2x3x4,x5[(x1,x2,x3,x4,x5)S].

This problem can be reduced to the previous one by a technique known as quantifier elimination and is therefore decidable as well; see Tarski (1951) and Seidenberg (1954).

The amount of computation required by the original methods of Tarski and Seidenberg was very high and more efficient methods have been devised since then. Consider a system of m polynomial (in)equalities, each of degree at most d, in n real variables. The problem of deciding whether the system has a solution can be solved with m(n+1)dO(n) arithmetic operations; see Basu, Pollack and Roy (1996). Similar bounds are possible when alternations of quantifiers are present (Basu et al., 1996); see also Blum et al. (1988, Section 18.6) and the references therein.

We now turn to undecidable problems. The usual technique for proving that a problem is undecidable involves the idea of a reduction. Suppose that some decision problem A can be reduced to problem B, in the sense that any algorithm for B leads to an algorithm for A. Suppose, furthermore, that A is known to be undecidable. It follows that B is undecidable as well. In this proof schema, we need a prototype problem whose undecidability can be established by some other means, and which can play the role of problem A. One such problem is the halting problem, described below. Its undecidability can be proved by a relatively simple "diagonalization" argument (Papadimitriou, 1994).

An instance of the halting problem consists of a description of a computing machine M, e.g., a particular RAM program, together with a particular input x to M. The question is whether M will eventually halt when started with input x. A possible approach to this problem might be to run M with input x and assert if it halts. But if M does not halt on input x, it is not apparent how to detect this in finite time and assert that the answer is . One could hope to infer the answer through some form of syntactic analysis of the program of machine M, but this is ruled out by the undecidability result.

A second example is provided by Hilbert's famous tenth problem on Diophantine equations, which is the following. We are given a polynomial in several variables, with integer coefficients, and we wish to decide whether it has an integer solution. Building on previous work by Davis and Robinson, this problem was proved undecidable in Matiyasevich (1970). An elementary exposition of Matiyasevich's Theorem can be found in Davis (1973).

A third example, which is remarkable in its simplicity, is Post's correspondence problem. We are given finitely many pairs of words2 (x1,y1),..., (xn,yn), and we wish to decide whether there exists some n1 and a finite sequence (i1,...,im) of integers in the range 1,...,n, such that

xi1xi2···xim=yi1yi2···yim.

This problem, proved undecidable in Post (1946), remains undecidable even if there are only seven pairs of words ( Matiyasevich & Sénizergues, 1996). The problem is decidable for two pairs of words ( Salomaa, 1985). The decidability of the intermediate cases (3n6) is unknown but is likely to be difficult to settle. The undecidability of the Post correspondence problem can be established by reducing the halting problem to it; for a proof see Hopcroft and Ullman (1969, Section 14.2).

2.4. Time complexity

If an algorithm halts, we define its running time to be the sum of the "costs" of each instruction carried out. Within the RAM model of computation, arithmetic operations involve a single instruction and could be assumed to have unit cost. Realistically, however, arithmetic operations take time that increases with the size (bit length) of the integers involved. This leads us to the bit model in which the cost of an arithmetic operation is taken to be the sum of the sizes (bit lengths) of the integers involved.

Of course, the running time of an algorithm will generally depend on the size of the instance. Furthermore, the running time can be different for different instances of the same size. We define the running time T(s) of an algorithm, as a function of size, to be the worst-case running time over all instances of size s.

We say that an algorithm runs in polynomial time if there exists some integer k such that

T(s)=O(sk)

and we define P as the class of all (decision) problems that admit polynomial-time algorithms. Due to a combination of practical and theoretical reasons, P is generally viewed as the class of problems that are efficiently solvable. It is most often the case that a computational problem can be efficiently solved in practice if and only if it belongs in P. The emphasis on worst-case performance is not fully satisfactory, as an algorithm may be fast on most instances but take exponential time only on a small minority of pathological instances. For example, the simplex method is known to have polynomial running time, on the average (Schrijver, 1986). However, it is often difficult to give a precise and reasonable meaning to the term "on the average". For this reason, the vast majority of complexity-theoretic research has focused on worst-case complexity.

Note that under the bit model that we have adopted, an algorithm that uses a polynomial number of arithmetic operations is not necessarily a polynomial-time algorithm. To appreciate this distinction, notice that the number 22n can be computed with n multiplications, by successive squaring, but this takes at least 2n steps (just to write the output). On the other hand, this distinction disappears for algorithms that produce intermediate and final results that are integers of size (bit length) bounded by a polynomial in the size of the original instance.

To make the above statement more concrete, let us consider the inversion of integer matrices. It can be carried out using Gaussian elimination, with O(n3) arithmetic operations, but more work is needed before asserting that we have a polynomial-time algorithm. The size of an instance involving an n×n matrix is O(n2 log U), where U is the magnitude of the largest entry of the matrix. It can be verified (see, e.g., Schrijver, 1986) that Gaussian elimination can be implemented with O(n3) arithmetic operations and so that every intermediate quantity produced is an integer of magnitude O((nU)nk ) for some integer k. Thus, intermediate quantities have size O(nk log (nU)). The cost of each arithmetic operation is bounded by that latter quantity, and the overall running time of the algorithm is O(nk+3 log (nU)). This is seen to be polynomial in the instance size O(n2 log U) and we conclude that matrix inversion can be done in polynomial time.

In the sequel, the bit model will always be in effect when talking about polynomial-time algorithms. On the other hand, unless there is an explicit statement to the contrary, we will provide complexity estimates in terms of arithmetic operations, which is less tedious.

2.5. Turing machines and other models

Turing machines provide a model of digital computation which is more primitive, hence harder to "program" than random access machines. However, their primitiveness becomes an advantage when they are manipulated for the purpose of proving theoretical results.

A Turing machine (TM, for short) uses a finite symbol alphabet that includes a "blank" symbol. It uses a linear tape (memory), which is an infinite sequence of cells, each one containing a symbol. It also uses a "head" that has a finite number of possible internal states and, at any time, scans a particular cell. When computation starts, the input is written on finitely many of the tape cells (the other cells are blank), the head is in a special "start" state, and it scans the first cell. At each time step, the machine reads the contents of the cell being scanned, and a transition is implemented. The effects of the transition are a function only of the current state and the contents of the scanned cell, and they result in (a) a new internal state, (b) a new symbol written in the scanned cell, and (c) a movement of the head one unit to the left or to the right. The computation terminates when the head enters a special "halting" state. At that point, the tape contents are interpreted as the output. A TM can be viewed as a representation of an algorithm, which is encoded in the transition function.

Most questions related to the operation of TMs are known to be undecidable (e.g., is a given TM guaranteed to halt on every input, does it halt on a particular input, etc.). Furthermore, the functions that can be computed by Turing machines are the same as those for the seemingly more powerful RAM model.

As far as time complexity is concerned, the most sensible notion of running time for TMs is to count the number of steps until the machine halts. Turing machines are capable of performing arithmetic operations on k-bit integers in time which is polynomial in k (verifying this is a simple but tedious exercise in low-level programming). Building on this observation, it can be shown that a TM can simulate a RAM with only polynomial slowdown. That is, if a RAM algorithm solves a certain problem in time O(sk) under the bit model, then there exists some l and an equivalent TM algorithm that takes time O(sk+l). The converse is also true because RAMs are at least as powerful as TMs.

This and similar observations have led to the sometimes called effective Church¯Turing thesis, which states that "any reasonable attempt to model mathematically computer algorithms and their time performance is bound to end up with a model of computation and associated time cost that is equivalent to Turing machines within a polynomial" (Papadimitriou, 1994).

2.6. Some polynomial-time solvable problems

We refer here to some important problems that are known to be polynomial-time solvable. We have already mentioned matrix inversion, and the same is true for problems such as solving systems of linear equations, computing determinants, or computing matrix rank. L (the problem of deciding whether a set of linear inequalities in real variables has a solution) can also be solved in polynomial time using interior point methods (Schrijver, 1986). The same is true for certain optimization problems such as convex quadratic programming (minimizing a convex quadratic function subject to linear equality and inequality constraints), or for solving linear matrix inequality (LMI) problems to within a prespecified error tolerance ( Boyd, El Ghaoui, Feron & Balakrishnan, 1994).

Closer to control theory, consider the problem of deciding whether a matrix A is stable, i.e., whether its spectral radius (A) (the maximum of the magnitudes of its eigenvalues) satisfies (A)<1. A solution to this problem follows from the following result: the matrix A is stable if and only if the Lyapunov equation APAT+I=P has a positive-definite solution P (Gantmacher, 1959). (In that case, the solution is unique.) Since the Lyapunov equation is linear in the unknown entries of P, we can compute a solution P (or decide it does not exist) in polynomial time. To check that P is positive definite, it suffices to compute the determinants of the n principal minors of P and verify that they are all positive. Since determinants can be computed in polynomial time, we have a polynomial-time solution to the stability problem. A similar polynomial-time solution is also possible for continuous-time stability (are all eigenvalues of A in the open-left half-plane?). The same questions about roots of polynomials can also be answered in polynomial time, since it is easy to construct a matrix A with a prespecified characteristic polynomial. An alternative method is to use the Routh test.

2.7. NP-completeness and beyond

There are many decidable problems of practical interest for which no polynomial-time algorithm is known, despite intensive research efforts. Many of these problems belong to a class known as NP (nondeterministic polynomial time), that includes all of P. A decision problem is said to belong to NP if every instance has a "certificate" of being a instance whose validity can be verified with a polynomial amount of computation. For example, consider the problem (). In this problem, we are given a number of zero¯one variables and a system of linear equality and inequality constraints, and we wish to determine whether a solution exists. If we have a instance, any feasible solution is a certificate that testifies to this and whose validity can be checked in polynomial time. (Simply, check that all the constraints are satisfied.)

The problem turns out to be a hardest problem within the class NP in the sense that every problem in NP can be reduced to , in polynomial time (Cook, 1971). Such problems are said to be NP-complete. More precisely, if A is any problem in NP, there is an algorithm that runs in polynomial time and which, given an instance I of problem A produces an "equivalent" instance I' of . Equivalence here means that I is a instance of A if and only if I' is a instance of .

If we had a polynomial-time algorithm for , then any problem in NP could also be solved in polynomial time by first reducing it to and then using a polynomial-time algorithm for . It would then follow that NP is the same as the class P of polynomial-time solvable problems. Whether this is the case or not is a major open problem in theoretical computer science. It is widely believed that PNP, although a proof is not in sight. If indeed PNP, then is not polynomial-time solvable, and the same is true for every other NP-complete problem. By now, NP-completeness has been established for thousands of problems (Karp, 1972; Garey & Johnson, 1979). We will have the opportunity of introducing several NP-complete problems later on. For now, let us mention , the problem of minimizing a quadratic ¯¯ but not necessarily convex ¯¯ cost function over a polyhedron.

If a problem A is at least as hard as some NP-complete problem B (in the sense that B can be reduced to A in polynomial time) we will use the term NP-hard to describe this situation. In particular, a decision problem is NP-complete if and only if it is NP-hard and belongs to NP.

There are also classes of problems that are much harder than NP-complete. For example, a problem is said to belong to exponential time (EXP) if it can be solved in time O(2sk ) for some k, where s is the instance size. Once more, EXP-complete problems are the hardest in this class and, in this case, they require provably exponential time (Papadimitriou, 1994).

Let us also define PSPACE, the class of all problems that can be solved by a Turing machine that uses a polynomial amount of memory (tape cells). We have NPPSPACE. (For example, can be solved with polynomial memory by simply checking all possible solutions.) However, it is not known whether the containment is proper. Given the present state of knowledge, it is even possible that PSPACE=P, but this is considered highly unlikely. If a problem is shown to be PSPACE-complete, this is viewed as a very strong indication that it is not polynomial-time solvable. To summarize, we have PNPPSPACEEXP. It is known that PEXP but it is unknown which of the above inclusions are strict.

2.8. Coping with NP-completeness

NP-hardness of a problem means that it is about as difficult as , and this is often interpreted as an indication of inherent intractability. Assuming that the conjecture PNP is true, an NP-hardness result eliminates the possibility of algorithms that run in polynomial time and that are correct for all problem instances. However, this is not reason enough for declaring the problem intractable and refraining from further research. Many NP-hard problems are routinely solved in practice, either exactly or approximately, using a variety of methods. For example, there are methods such as branch-and-bound that require exponential time in the worst case, but run fairly fast on many problems and instances of practical interest (Nemhauser & Wolsey, 1988).

Another point that needs to be appreciated is that not all NP-complete problems are equally hard. A standard example is provided by the 0¯1 problem (see, e.g., Bertsimas & Tsitsiklis, 1997)

where the data ci, wi, and K are nonnegative integers. The problem is NP-hard, but can be solved in pseudo-polynomial time, with n2 maxi|ci| arithmetic operations. This is not polynomial time, because the instance size is of the order of log K+ilog(ciwi), and the running time is an exponential function of the instance size. However, unless the numbers ci are extremely large, this can be a practically viable algorithm.

In addition, many NP-complete (or NP-hard) problems become much easier if we are willing to settle for approximate solutions. For example, for any >0, we can get a feasible solution of the 0¯1 problem, whose value is no less than (1-) times the optimal value, and this can be done in polynomial time, with O(n3/) arithmetic operations. More generally, a minimization (respectively, maximization) problem is said to admit a polynomial-time approximation scheme if for every >0, there exists an algorithm (depending on ) whose running time is polynomial in the instance size, and which outputs a solution which comes within a factor of 1+ (respectively, 1-) from the optimum. However, some NP-hard problems are not amenable to approximate solution, and there is a rich theory pertaining to this matter (Papadimitriou, 1994). We revisit this subject in Section 3.2.

Finally, let us mention the possibility of using polynomial-time randomized algorithms that provide solutions to arbitrarily high levels of accuracy and confidence (see Tempo et al. (1996), Khargonekar and Tikku (1996), Vidyasagar (1998) and Vidyasagar and Blondel (1999) for a description of such methods applied to control problems).

3. Linear systems

Many questions about linear systems are decidable, and can be decided efficiently. For example, controllability, stability, and observability of continuous- or discrete-time linear systems can be decided in polynomial time. One problem for which no algorithm is yet available is the "hyperplane hitting problem" which arises, among other situations, in sliding mode control. In this problem, we are given a discrete-time autonomous system xt+1=Axt, together with the initial state x0, and we wish to determine whether the state xt eventually lands on the hyperplane cTx=0, that is, whether we have cTAtx0=0 for some t0. This problem is the system-theoretic transcription of a well-known problem in the theory of formal power series (Salomaa & Soittola, 1978). It is also a long-standing open problem in number theory. The problem is proved NP-hard in Blondel and Portier (1999a) but it is unknown whether it is decidable. The present consensus among number theorists is that an algorithm should exist. This has been proved to be the case for a large number of special cases, including the generic situation where the eigenvalues of A are all distinct; see Mignotte, Shorey and Tijdeman (1984), and also Samake (1996) for recent extensions.

3.1. Stable and unstable polynomials in polynomial families

Many computational complexity results in control theory arise when considering families of polynomials or matrices. The interest in the stability analysis of such families emerged in the early 1980s after the popularization of Kharitonov's theorem in the control engineering literature. In the following, a polynomial is said to be stable if all of its roots have negative real parts.3

In Kharitonov (1978), it is proved that every real polynomial in the interval family of polynomials

{a0+a1x+a2x2+···+anxn: ai-aiai+,i=0,...,n}

is stable if and only if four special members of the family with extreme coefficients (the Kharitonov polynomials) are stable.

The original proof of the theorem is in Russian. Simple geometrical proofs can be found in Dasgupta (1988) and Minnichelli, Anagnost and Desoer (1989). Using Kharitonov's theorem, it is easy to design a polynomial-time algorithm that decides whether a given interval family is stable: it suffices to apply stability tests to the four Kharitonov polynomials. The relevance of this result to robust stability analysis provided motivation for various attempts at generalizations but with limited success. This experience opened up the possibility that broad generalizations of Kharitonov's Theorem, of comparable simplicity, do not exist. Indeed, results from computational complexity show that, unless P=NP, a variety of other types of polynomial families do not admit stability criteria that can be tested efficiently. We review a number of such results.

Interval polynomial families are special cases of affine polynomial families. Let pi(x), i=1,...,n, be real polynomials. An affine polynomial family is a set of polynomials of the form

{p0(x)+a1p1(x)+a2p2(x)+···+ampm(x):ai-aiai+, i=1,...,m}.

An edge of the family is a subset (i.e., a set of polynomials) for which all of the parameters ai are fixed to their upper or lower bounds ai+ or ai-, with the exception of one parameter, which is allowed to vary within the corresponding interval. It is known (Bartlett, Hollot & Huang, 1988) that the stability of an affine polynomial family is equivalent to the stability of all of its edges. Furthermore, the stability of an edge can be checked in polynomial time by using an edge stability condition given in Bialas (1985). There are m2m-1 edges in an affine polynomial family with m parameters. The elementary stability checking procedure that consists in checking all edges leads thus to a number of operations that is exponential in the number of parameters. This procedure can be improved by identifying critical edges. It is shown in Sideris (1991) that the stability of an affine polynomial family of degree n, with m parameters, can be checked in polynomial time, with O(m3n2) arithmetic operations.

Broader classes of polynomial families can be constructed as follows. A multilinear polynomial family (or, more precisely, a multiaffine polynomial family) is a set of polynomials of the form

{b0(a0,...,am)+b1(a0,...,am)x+b2(a0,...,am)x2+···+bn(a0,...,am)xn:ai-aiai+, i=0,...m},

where each coefficient bi(a0,...,an) is an affine function of each parameter ai when the remaining parameters are held fixed. This structure arises, for example, when one considers the characteristic polynomial of an interval family of matrices (see the next section). Several contributions have dealt with the stability of such families. In contrast to the affine case, however, no polynomial-time stability checking algorithms are available. Indeed, using Proposition 2.1 in Nemirovskii (1993), it is easily concluded that, unless P=NP, no such algorithm exists. We present here an elementary proof of an improved version of this result. We define a bilinear polynomial family as a multilinear polynomial family in which each coefficient function depends only on two parameters. We show that, unless P=NP, the stability of bilinear polynomial families of degree one is not decidable in polynomial time. We provide a complete proof because of its simplicity and because it adequately illustrates the usual way that such negative results are proved. The proof is based on a reduction of the 2- problem, which is known to be NP-complete (Garey & Johnson, 1979).

M 2-

Instance: A set U={u1,...,um} of Boolean variables, a collection C of k clauses over U, such that there are at most two literals in each clause, and a positive integer Kk.4

Question: Is there a truth assignment for the variables in U that simultaneously satisfies at least K of the clauses in C?The stability of bilinear polynomial families of degree one is NP-hard to decide.Given an arbitrary instance of 2-, we will construct an equivalent instance of the stability problem. Any clause cC is of one of the following forms: c=ui, , c=uiuj, , or , for some i,j. To c=ui we associate the affine function ai, to we associate 1-ai, to c=uiuj we associate ai+aj-aiaj, to we associate 1-aj+aiaj, and to we associate 1-aiaj.Let A(a1,...,an) be the sum of the polynomials associated with the clauses in C and consider the bilinear polynomial family

{x+(K-A(a1,...,an)): 0ai1}.

All polynomials in the family are stable if and only if the root A(a1,...,an)-K is negative for all choices of the ai in [0,1]. That is, if and only if the maximum of A(a1,...,an) is less than K. Note that the maximum of A(a1,...,an) for ai[0,1] is attained for ai{0,1}, and is therefore equal to the maximum (over all possible truth assignments) number of satisfied clauses. Thus, we have a stable family if and only if we have a instance of 2-. We therefore have a polynomial time reduction of an NP-complete problem to the problem of interest, and the result follows.

We have seen that the stability of affine polynomial families can be checked in polynomial time, whereas the stability of bilinear polynomial families is NP-hard. Thus, the tractability borderline is somewhere between affine and bilinear families. Unless P=NP, efficient stability checking algorithms can only be derived for particular multilinear polynomial families, see, e.g., Fu, Dasgupta and Blondel (1995).

The problem of deciding whether a given family of polynomials is stable is equivalent to the problem of deciding whether the family contains an unstable polynomial. Both problems are analysis problems. A related problem is that of asking whether a family contains a stable polynomial, which is the same as asking whether all polynomials in the family are unstable. This is a design problem and, although its complexity is related to the complexity of the analysis problem, the relation is not as immediate as it may first seem.

Kharitonov's theorem does not hold for instability. In general, it is not true that instability of all four Kharitonov polynomials of an interval family of polynomials ensures instability of the entire family. Similarly, it is not true in general that instability of all edges of an affine polynomial family ensures instability of the entire family. Existence of a polynomial time algorithm for checking the presence of a stable polynomial in an affine polynomial family is an open problem. This problem is of interest because it is intimately related to basic control design problems. On the other hand, the proof given above can be easily adapted to prove that the problem of deciding whether there is a stable polynomial in a given bilinear family of polynomials is NP-hard.

3.2. Stability of matrix families

If A- and A+ are square matrices of the same dimensions, we define the interval matrix family [A-,A+] as the set of all matrices A that satisfy the inequality A-AA+ componentwise, i.e., (A-)ijAij(A+)ij for every i and j.

For interval matrix families, there are no results analogous to Kharitonov's theorem. In fact, the stability of all edges of an interval matrix family does not necessarily imply the stability of the entire family. There are however some results, for the symmetric case, in terms of the extreme points of the family. An interval matrix family [A-,A+] is said to be symmetric if the matrices A- and A+ are symmetric. (Note, however, that in general, a symmetric interval matrix family [A-,A+] also contains nonsymmetric matrices.) Necessary and sufficient conditions for stability of symmetric interval matrices, formulated in terms of the stability of a finite subset of matrices in the family, are given in Soh (1990), Hertz (1992), and Rohn (1994a). Still, in all of these references, the number of matrices that have to be tested is exponential in the dimension of the matrices. This seems to be unavoidable given the negative results that we describe next.

The computational complexity analysis of problems associated with interval matrix families was initiated in Poljak and Rohn (1993) and Nemirovskii (1993). 5 A combination of the results presented in these two papers is as follows. An elegant and economical presentation is provided in Rohn (1994b).

For interval matrix families, the following four problems are NP-hard:


The NP-hardness proof of (i) given by Poljak and Rohn involves the problem, which is defined as follows. We are given an undirected graph, and we wish to partition the set of nodes into two disjoint subsets, in a way that maximizes the number of edges with one endpoint in each subset. The decision version of this problem ("Is the maximum larger than a given number?") is NP-complete (Papadimitriou, 1994; Poljak & Rohn, 1993) shows that can be reduced to . Furthermore, remains NP-hard even if the size (bit length) of the entries of the n×n matrices A- and A+ is O(log n). Problems with such a property, that remain NP-hard even if the size of each number in the input is restricted to be small (of the order of a polynomial in the logarithm of the overall instance size), are said to be strongly NP-hard. The complexity of such problems is of an inherently combinatorial nature, rather than being due to excessively long numerical input.

We can rephrase the decision problem as an optimization problem, and look at the possibility of approximations. The center of the interval matrix family [A-,A+] is Ac=(A++A-)/2 and its error is Ae=(A+-A-)/2. The interval matrix family [Ac-Ae,Ac+Ae] is equal to {Ac} when =0, and is equal to the interval family [A-,A+] when =1. The radius of stability (Ac,Ae) associated with Ac and Ae is the smallest value of for which [Ac-Ae,Ac+Ae] contains a singular matrix. Clearly, all matrices in [A-,A+] are nonsingular iff (Ac,Ae)>1 and, therefore, deciding whether (Ac,Ae)>1 is also NP-hard. However, a stronger result that refers to the impossibility of approximating (Ac,Ae) is also possible, as we discuss next.

It is known that unless P=NP, does not admit a polynomial-time approximation scheme as defined in Section 2.8 ( Papadimitriou, 1994). It follows that the same is true for the optimization version of (Coxson & De Marco, 1994); see also Demmel (1992).

The NP-hardness of the other three problems in the above theorem was shown in Nemirovskii (1993), by using a reduction of the classical problem,6 which is NP-complete (Garey & Johnson, 1979). The reductions produce very particular classes of interval matrix families. Thus, restricted versions of the problems that encompass these classes are also NP-hard. For example, is NP-hard when all the entries of the matrices in the family are fixed, except for some of the entries of a single row and a single column, which are allowed to vary between -1 and 1.

Even though is NP-complete, it can be solved in time polynomial in the number of integers and exponential in their size (bit length). (In fact, it is a special case of the 0¯1 problem discussed in Section 2.8.) Thus, the proof in Nemirovskii (1993) still leaves the possibility that problems such as can be solved in time polynomial in the dimensions of the matrices, when their entries have small bit length, e.g., logarithmic in the dimension of the matrices. However, this possibility is ruled out by adapting Nemirovskii's proof with the help of the reduction used in Poljak and Rohn (1993), and all of these problems are strongly NP-complete ( Coxson, 1993).

An interval matrix family can always be expressed in the form

{A0+a1A1+···+amAm: 0ai1, 1im},

where each matrix Ai has a single nonzero entry. A rank- affine matrix family is a family of the form above, where the matrices Ai have rank that is less than or equal to . The stability of rank-1 affine families is NP-hard to decide, and the special case m=1 is solvable in polynomial time without any rank restrictions; see Fu and Barmish (1988). Besides these two special cases, the territory is unmarked. Exponential time branch-and-bound algorithms for the case of arbitrary m and are proposed in Kokame and Mori (1992). A solution in terms of finitely many linear programs is proposed in Barmish, Floudas, Hollot and Tempo (1994).

There are many more results on the complexity of interval linear algebra problems. A survey of some of these results is given in Kreinovich, Lakeyev, Rohn and Kahl (1995) and Rohn (1997). See also the recent book ( Kreinovich et al., 1997).

Let us finally mention a different problem but with a somewhat similar flavor. Given a system with delays, of the form

we wish to determine whether the system is stable for all allowed values of the delays hk. It turns out that this problem is NP-hard for the case where each hk is constrained to lie within an interval [hk-,hk+], as well as for the unconstrained case, where hk[0,+) (Toker & Ozbay, 1996).

3.3. Structured singular values

The results in the preceding subsection have direct implications for the computation of the structured singular value of a matrix, a fact that was hinted at in Poljak and Rohn (1993). The structured singular value generalizes the notion of singular value and arises in many robust control problems involving systems with uncertain parameters.

We start with a square, possibly complex, matrix M of dimensions n×n and we consider a set of block-diagonal perturbation matrices , of the form =diag{1,..., k}, where each i has dimensions ni×ni and n1+···+nk=n. Each diagonal block is constrained to be of a particular type, e.g., a real multiple of the identity matrix, a complex multiple of the identity matrix, or an arbitrary real (or complex) matrix. Having defined the set of matrices in terms of such constraints, the corresponding structured singular value is defined by

where stands for the largest singular value of . When , (M) is equal to the spectral radius of M, and when , (M) is equal to the maximum singular value of M. For an introduction to structured singular values and their importance in system analysis and design, see Doyle (1982), Safonov (1982) and Zhou, Doyle and Glover (1995).

Many researchers had worked on algorithms for computing exactly, under various uncertainty structures, until it was established that the problem of deciding whether (M)1 is NP-hard for the following cases:


The result for the first two cases is a corollary of the results in Poljak and Rohn (1993). It is also derived in Braatz, Young, Doyle and Morari (1994), by showing that quadratic programming can be reduced to the problem of computing . The third result was shown later inToker and Ozbay (1998) by establishing that a suitably defined quadratic programming problem involving complex variables is NP-hard, and then reducing it to the purely complex problem.

Partly as a consequence of such results, research efforts have shifted to the problem of finding approximations of which are easy to compute and that are as close as possible to the exact value of ; see Fu (1999) and Toker and de Jager (1999) for some discussion on computational complexity problems related to the approximate computation of , and Trei (1999) for a negative result on an approach that had seemed promising.

Finally, let us mention that Fu and Dasgupta (1998) have studied the complexity of computing the real structured singular value with perturbations measured by a p-norm. For a set of block diagonal matrices where each block i is a real multiple iI of the identity, define

where =(1,..., k) and =diag{1I,..., kI}. The quantity p coincides with when p= and so the problem of deciding whether (M)1 is NP-hard. Fu and Dasgupta show that the problem of deciding whether p(M)1 is NP-hard for any p[1,] when the size of the smallest block exceeds one. For blocks of size one and finite p, the question of NP-hardness is open.

3.4. Existence of a stable matrix in an interval family; static output feedback

We now turn our attention to design, rather than analysis, questions for interval matrix families, and consider the following problem.

Instance: An interval family of matrices.

Question: Does the family contain a stable matrix?

This problem is NP-hard. It remains so even if the interval family is symmetric and all the entries of the family are fixed, except for some of the entries of a single row and a single column, which can take values between -1 and 1. The proof is given in Blondel and Tsitsiklis (1997) and involves again a reduction of the problem. In contrast to the corresponding analysis question discussed in Section 3.2, it is not known whether is NP-hard in the strong sense. A variation of is the problem in which we only look at symmetric matrices. Consider the semidefinite programming problem of minimizing subject to the constraint that I-A is nonnegative definite and subject to the interval constraints on the symmetric matrix A.7 Stability is then equivalent to the optimal value of being negative. This semidefinite programming problem can be solved in polynomial time to any fixed desired accuracy. However, the complexity of an exact solution is unknown.

This problem bears some similarities to the widely studied, and still unsolved, static output feedback problem. We are given matrices A, B, and C, and we are interested in determining whether there exists a "gain" matrix K such that A+BKC is stable. Despite numerous contributions since the early 1960s, a satisfactory answer to this problem has yet to be found. The problem is often cited as one of the difficult open problems in systems and control (Blondel, Gevers & Lindquist, 1995). See Syrmos, Abdallah and Dorato (1994) for a survey, and Rosenthal and Wang (1997) for genericity results on the related pole assignment problem. From a complexity point of view, the problem is algorithmically decidable using Tarski elimination, as pointed out in Anderson, Bose and Jury (1975).

Still, despite various attempts, it is unclear whether the problem is NP-hard. In Fu and Luo (1997), the authors show that a certain type of mixed LMI can be used to solve the static output feedback problem, but they then show that the proposed mixed LMI is NP-hard.

Using the NP-hardness result for , it is easily shown that the static output feedback problem is NP-hard if we also require the gain matrix K to satisfy interval constraints. However, the relevance of this observation is unclear, because the problem with interval constraints remains NP-hard for the special case of state feedback (given A and B, determine whether there exists some K within a given interval family such that A+BK is stable), even though it can be solved in polynomial time in the absence of interval constraints (Blondel & Tsitsiklis, 1997a). Let us also mention some related problems that have been shown to be NP-hard in the same reference. These are the problems of simultaneous stabilization by output feedback (see also Toker and Ozbay, 1995), decentralized output feedback stabilization using a norm bounded controller, and decentralized stabilization using identical controllers.

3.5. Time-varying systems

We have been discussing so far problems related to uncertain systems but with time-invariant uncertain parameters. We now turn to a discussion of problems relevant to systems with time-varying uncertainty, and we will be focusing on discrete-time systems.

We consider a time-varying system of the form xt+1=Atxt, where is a finite set of matrices, and At for every t0. We do not impose any restrictions on the sequence of matrices At. Starting from the initial state x0, we obtain

xt+1=At···A1A0x0.

We wish to characterize how xt grows with t. For this purpose, let ||·|| be an induced matrix norm. The joint spectral radius is defined in Rota and Strang (1960) by

where

for k1. (See also Berger and Wang (1992) for another characterization of in terms of the spectral radius of the matrix products.) It turns out that the limit always exists and the value of is the same for every induced matrix norm. The same comment applies to the definitions of and P() given later in this section.

For a given finite set of matrices , we can now pose questions such as the following:

Do we have ?

Are all possible matrix products Ak···A1 stable?

Do we have ?

Do all infinite matrix products converge to zero?

Do all periodic matrix products converge to zero?

Do we have xt0 for every x0 and any possible sequence of matrices?

Do we have xt0 for every x0 and any possible periodic sequence of matrices?

Is the set of all possible matrix products Ak···A1 bounded?

The above questions are not equivalent, but they are closely related. Due to their relevance in many practical settings, they have been extensively studied in recent years. For example, certain inequalities proved in Daubechies and Lagarias (1992, Lemma 3.1) can be used to derive algorithms that compute arbitrarily precise approximations for ; see Gripenberg (1996) for one such algorithm. These approximation algorithms can then be used in procedures that decide, after finitely many steps, whether or . Related procedures are given in Brayton and Tong (1980) and Barabanov (1988). The drawback of all of these procedures is that they need not terminate when happens to be equal to 1. In fact, the ability to compute arbitrarily accurate approximations of does not rule out the possibility that the problem of deciding whether is undecidable, and it is so far unknown whether this is the case or not; see Lagarias and Wang (1995) for a discussion of this issue and for a description of its connection with the so-called finiteness conjecture, 8 as well as the discussions in Gurvits (1995) and the NP-hardness result in Toker (1997). It turns out however that the slightly different problem of deciding whether (as well as the question of boundedness of the set of all matrix products) is undecidable (Blondel and Tsitsiklis, 1999b). Both problems remain undecidable in the case where contains only two matrices. Another negative result is given in Kozyakin (1990), which establishes that several questions of the type introduced in this section cannot be answered using a restricted class of algebraic algorithms, because the set of all instances has infinitely many connected components, even for pairs of 2×2 matrices.

As a practical matter, efficient approximation algorithms for computing might suffice. However, the results of Tsitsiklis and Blondel (1997a, Theorem 1) place some limitations on possible approximation algorithms. In particular, it is shown that unless P=NP, there is no algorithm that can compute with percent relative accuracy and which runs in time polynomial in log(1/) and the size of .9 Furthermore, this negative result continues to hold even if we restrict the set to consist of only two matrices with 0¯1 entries. Despite this negative result it is still conceivable that for any fixed >0, there exists a polynomial-time algorithm that computes with relative accuracy , but it is not known whether this is the case or not.

If a polynomial-time algorithm were available for checking the stability of all products of two given matrices, then the algorithm could be used in conjunction with binary search to approximate the joint spectral radius in polynomial time. As a consequence, it is NP-hard to decide whether all products of two given matrices (with rational entries) are stable. This is true even if all nonzero entries of the two matrices are constrained to be equal.10 The decidability of this problem is not known. On the other hand, polynomial-time stability checking algorithms are available when all nonzero entries are equal to one; see Gurvits (1996).

Similar to the joint spectral radius, we can define the lower spectral radius by

where

for k1.

The joint spectral radius measures the maximal average norm of long products of matrices taken from a finite set, whereas the lower spectral radius measures the minimal average norm. The problem of computing is intimately related to the mortality problem. We say that the set of matrices is mortal if the zero matrix can be expressed as the product of finitely many matrices from . Reference Paterson (1970) establishes a relation between word concatenation and matrix multiplication, and reduces Post's correspondence problem to the problem of deciding whether is mortal. In particular, it shows that the mortality problem is undecidable, even if consists of only 2np+2 integer matrices of dimensions 3×3, as long as Post's correspondence problem with np pairs of words is undecidable. (As discussed in Section 2.3, we can take np=7.) Building on this result (Blondel & Tsitsiklis, 1997b) establishes that the mortality problem remains undecidable even if consists of only two integer matrices of dimensions (6np+6)×(6np+6). The number of 3×3 integer matrices involved in Paterson's proof can be reduced to np+2 by considering the modified Post correspondence problem in which the initial pair of words used in the correspondence is predetermined and is part of the instance description (Hopcroft & Ullman, 1969); see Bournez and Branicky (1998) for a proof. An easy argument shows that, if mortality of d matrices of size n×n is undecidable, then the same applies to pairs of matrices of size nd×nd (Blondel & Tsitsiklis, 1997b). Thus, mortality of two integer matrices of size 3(np+2) is undecidable. Moreover, all 3×3 matrices involved are diagonalizable, and so, by adapting the argument in Blondel and Tsitsiklis (1997b), one can conclude that the mortality problem for pairs of 3(np+2) integer matrices, one of which is diagonal, is also undecidable. For the special cases of matrices with nonnegative entries, or when we restrict to products of the form A1···Ak for fixed k, the mortality problem can be solved (in exponential time), but is actually NP-hard (Blondel & Tsitsiklis, 1997b). See also Bournez and Branicky (1998) for a discussion of the mortality problem for 2×2 matrices, and Krom (1981), Krom and Krom (1989) and Klarner, Birget and Satterfield (1991) for more results on problems of similar flavor.

As long as the matrices have integer entries, we have either (which is the case if and only if is mortal) or . Thus the undecidability results for the mortality problem preclude the possibility of an exact, or even approximate computation of (Blondel & Tsitsiklis, 1997b). By the same argument, the problem of determining whether some product of two given matrices A0 and A1 is stable is also undecidable. Building on the comments made in the previous paragraph, it can be shown that this is true even if the two matrices have integer entries, are of size 27×27, and one of them is diagonal. On the other hand, if it is a priori known that , the question can be decided in polynomial time (Gurvits, 1996).

While and refer to two extreme ways of choosing a sequence of matrices, we can define an intermediate quantity if we let the matrices At be generated randomly and independently, according to a probability distribution P on the set . The largest Lyapunov exponent associated with P and is then defined by

where E[·] stands for expectation. We define the Lyapunov spectral radius P() by

P()=e(,P)

and it is easily verified that

Similar to our earlier arguments, the undecidability of mortality can be used to establish that P() cannot be computed exactly or approximately (Blondel & Tsitsiklis, 1997b).

It is interesting to contrast the latter result with what is known for analogous quantities defined in the max-plus algebra, where the usual operations of addition and multiplication are replaced by taking the maximum of two numbers, and addition, respectively (Baccelli, Cohen, Olsder & Quadrat, 1992). Products of matrices in the max-plus algebra correspond to performance measures for discrete-event systems, and the Lyapunov spectral radius is particularly interesting. In this context, the joint spectral radius can be computed in polynomial time, but the question whether <1 is NP-hard for both the lower spectral radius and for a quantity analogous to the Lyapunov spectral radius (Blondel, Gaubert & Tsitsiklis, 1998). In terms of computability, the latter quantity can be approximated (in exponential time) to any degree of accuracy ( Baccelli et al., 1992), but it is unknown whether the question "is <1?" is decidable.

4. Nonlinear and hybrid systems

4.1. Introduction

Control problems are in general easier to solve for linear systems than for nonlinear ones. Even for systems that seem mildly nonlinear, simply stated control questions may become intractable or, in many cases, undecidable. We review in this section some of the available complexity results, including related results for hybrid systems. We look at four particular control problems.

S

Input: A system xt+1=f (xt,ut) and an initial state x0.

Question: Does there exist some T1 and controls ut, t=0,...,T-1, such that xT=0?

N (All states can be driven to the origin.)

Input: A system xt+1=f (xt,ut).

Question: Is it true that for every initial state x0, there exists some T1 and controls ut, t=0,...,T-1, such that xT=0?

T

Input: A system xt+1=f (xt) and an initial state x0.

Question: Does there exist some T1 such that xT=0?

A

Input: A system xt+1=f (xt).

Question: Is it true that for every x0 there exists some T1 such that xT=0?

Asymptotic versions of these definitions are possible by requiring the sequences to converge to the given state rather than reaching it in finite time. For example, in a problem , we have an autonomous system and the objective is to decide whether xt0 for any initial state x0. The first two problems are control design problems, whereas the last two are analysis problems. When algorithms exist for the design problems, they can be used for the corresponding analysis problem as well. Thus, analysis problems are in general easier to solve. When looking for negative complexity-theoretic results, the reverse is true; analysis problems are harder to show undecidable or NP-hard than control problems.

In all of these problems, we need to assume that the system is "given" in terms of a finite-length description of the function f. For example, if f is polynomial with rational coefficients, it is "given" in terms of its coefficients. For linear systems the above questions are all decidable in polynomial time. On the other hand, no algorithms exist for general nonlinear systems. Stated at this level of generality, these problems are not interesting, because they are far too difficult. For example, as pointed in Sontag (1995), the null-controllability question for general nonlinear systems includes the problem of deciding whether a given arbitrary nonlinear equation (u)=0 has a solution. Indeed, for a given function , consider the system xt+1=(ut), which is null controllable if and only if has a zero. Thus, the null-controllability question for nonlinear systems is at least as hard as deciding whether a given nonlinear function has a zero, which is far too general a problem. For a problem to be interesting, we need to constrain the type of nonlinear systems considered. In the subsections that follow we consider several classes of systems, such as systems with a single nonlinearity, systems with componentwise nonlinearities, piecewise-linear systems, and hybrid systems.

4.2. Systems with a single nonlinearity

Even for systems that involve a single scalar nonlinearity, algorithms for deciding stability are inherently inefficient. To illustrate this, let us fix an arbitrary nonconstant scalar function : RR that satisfies

(2)

Having fixed , let us consider the problem in which we are given A0,A1Qn×n, and cQn as input, and we wish to determine whether all trajectories of the system

xt+1=(A0+(cTxt)A1)xt(3)

converge to the origin. If were constant, system (3) would be linear and this question could be decided in polynomial time. On the other hand, the problem is known to be NP-hard for any fixed nonconstant that satisfies (2) ( Blondel & Tsitsiklis, 1999a).

If we focus on the particular case where is the sign function, we obtain an NP-hardness result for a very simple class of nonlinear systems, namely systems that are linear on each side of a hyperplane:

(4)

Such systems could arise when a linear system is controlled using a "bang¯bang" controller. They also represent one of the simplest types of hybrid systems.

It is not clear whether questions related to the stability of systems (3) are actually decidable. We do not know of any nonconstant function for which has been proved decidable or undecidable.

System (4) is of the form xt+1=Atxt where each At belongs to the set {A-,A+}. This is a situation similar to the one considered in Section 3.5, except that now the sequence of matrices is not arbitrary, but is determined by the state sequence. Still, this does not help in reducing complexity as we now argue using an example taken from Blondel and Tsitsiklis (1998a). Consider a system described by a state vector (vt,yt,zt), where vt and yt are scalars and zt is a vector in Rn, and the dynamics is of the form

and

This system consists of two linear systems, each of which is enabled in one of two half-spaces, as determined by the sign of yt. Given that y0 can be any real number, it can be verified that the sequence sign(yt) is completely arbitrary, which then implies that the matrices A- and A+ can be multiplied in an arbitrary order. Clearly, all trajectories converge to zero if and only if zt always converges to zero and the problem is at least as hard the corresponding question from Section 3.5 in which arbitrary products were allowed. If the latter problem is ever shown to be undecidable, the same will be automatically true for the problem considered here. On the other hand, since the boundedness of all products of A- and A+ is undecidable (Blondel & Tsitsiklis, 1999b), it follows that the uniform boundedness of the trajectories of system (4), for all initial states in a bounded set, is also undecidable.

One can easily adapt (3) to include the possibility of a control action, as in a system of the form

xt+1=(A0+(cTxt)A1)xt+But.(5)

The question of for system (5), subsumes the question of whether the autonomous system (3) ¯¯ just consider the special case where B=0 ¯¯ and NP-hardness follows. One can in fact say more than that. When is the sign function, then and for systems (5) becomes undecidable ( Blondel & Tsitsiklis, 1999a, Theorem 2).

Let us finally note that for functions that have finite range, systems (5) become piecewise linear. We shall say more about the complexity of such systems in Section 4.4.

4.3. Systems with componentwise nonlinearities

Let us fix a scalar function : RR, and consider systems of the form

xt+1=(Axt),(6)

where the matrix A has size n×n and is applied componentwise, i.e.,

(q1,...,qn)=((q1),..., (qn)).

Systems of this type arise in a wide variety of situations. When is linear, we have a linear system and most dynamical properties of interest can be checked efficiently. When has finite range, the entries of the state vector take values in a finite set, the system evolves on a finite state space after the first time step, and dynamical properties can be decided in time which is polynomial in the number of possible states. Note, however, that the number of states generally increases exponentially in n. If is the sign function and A is symmetric, system (6) is guaranteed to have a fixed point, but it is not known whether it can be found in polynomial time ( Papadimitriou, 1994).

Recurrent artificial neural networks are commonly of the form (6) ( Sontag, 1996b) and, in that context, is referred to as the activation function. Some of the activation functions that are common in the neural network literature are the logistic function (x)=1/(1+e-x), the trigonometric function (x)=arctan(x), and the saturated linear function

Note that all of these functions are continuous and have finite limits on both ends of the real axis, which is a common assumption in the context of artificial neural networks. Other motivations for studying systems of type (6) arise in the context of fixed-point digital filters and of linear systems with saturation limits on the state or control variables. See the book Liu and Michel (1994) for further motivation and many references related to neural networks and filter design. Finally, we will also discuss the cut function, defined by

which is probably the simplest nonlinear but piecewise linear function.

It turns out that systems of the form (6) have vastly different characteristics from linear systems. A result announced in Siegelmann and Sontag (1991) and completed in Siegelmann and Sontag (1995), shows that saturated linear systems are capable of simulating Turing machines. This is done by suitably encoding the transition rules of the Turing machine in the matrix A, while the tape contents and the machine's internal state are encoded on some of the states of the saturated linear system. Thus, as computational devices, linear saturated systems are as powerful as Turing machines. This abstract result has several system-theoretic implications. We describe a representative one. Recall that the halting problem is undecidable for Turing machines. Therefore, the problem of deciding whether a given initial state of a saturated linear system eventually leads to a state that encodes a halting configuration, is also undecidable. This halting state can be chosen to be the origin (see Sontag, 1995) and it follows that is undecidable for saturated linear systems.

By using a universal Turing machine,11 one can in fact prove that for saturated linear systems, is undecidable for some particular matrix A. More concretely, there exists a particular integer matrix A (of size approximately equal to 1000×1000) for which there exists no algorithm that takes an initial state x0 (with rational components) as input and decides whether the state will eventually hit the origin (Siegelmann & Sontag, 1995). The problem of finding the smallest possible such matrix is a difficult open problem that is related to the smallest possible size of universal Turing machines ( Rogozhin, 1996).

The initial result by Siegelmann and Sontag has prompted efforts to understand the necessary properties of a function , under which Turing machine simulation is possible. The fact that such simulations are possible with the cut function is proved in an elementary and simple way in Hyotyniemu (1997). 12 In Koiran (1996), it is shown that Turing machines can be simulated if eventually becomes constant on both ends of the real line and is twice differentiable with nonzero derivative on some open interval. Note, however, that the function =arctan and other standard functions in the neural network literature do not satisfy these hypothesis. The conditions are further relaxed in Kilian and Siegelmann (1996). The authors offer a sketch of a proof that Turing machines can be simulated for any function in a certain class that includes, among others, the functions described above and all the functions that are classically used in artificial neural networks. The functions do not need to become ultimately constant but need to be monotone.

Using an argument similar to the one provided earlier for saturated linear systems, we arrive at the conclusion that for systems (6) is undecidable when is the saturated linear function, the cut function, the logistic function, or any function that belongs to the classes defined in Koiran (1996) and Kilian and Siegelmann (1996). The undecidability of immediately implies the same for for systems of the form

xt+1=(Axt+But).(7)

These results do not have any direct implications for the problem . For saturated linear systems, this problem was conjectured to be undecidable by Sontag in Sontag (1995). A proof that this is indeed the case is given in Blondel, Bournez, Koiran and Tsitsiklis (1999b) and establishes the undecidability of several related global properties such as . The main idea is to simulate a Turing machine by a saturated linear system so that all trajectories converge to the origin if and only if the Turing machine halts for every initial configuration. By a result of Hooper (Hooper, 1966), the latter Turing machine problem is undecidable, hence the proof of Sontag's conjecture.

4.4. Piecewise linear and hybrid systems

Hybrid systems involve a combination of continuous dynamics (e.g., differential or difference equations) and discrete dynamics. There is no consensus on a single general model of hybrid systems, but many discrete-time models include the class that we describe next.

We start with a partition of Rn into finitely many disjoint subsets H1,H2,...,Hm. We assume that a different linear system is associated with each subset, and the overall system is described by

xt+1=Aixt when xtHi.(8)

When the subsets Hi are defined in terms of a finite number of linear inequalities, the systems of the form (8) are the piecewise linear systems introduced by Sontag (1981) as a unifying model for describing interconnections between automata and linear systems. See Sontag (1985) for a characterization of the complexity of questions related to fixed time horizon properties of such systems, and Sontag (1996a) for a more recent account of available results.

Note that systems with componentwise nonlinearities, and with equal to the saturated linear function, are a special case of piecewise linear systems. Hence, using the results of the previous section, we deduce that piecewise linear systems can simulate Turing machines and is undecidable for piecewise linear systems. It is known that simulation of Turing machines is possible by using piecewise affine systems in state dimension two (Moore, 1990), but not in dimension one ( Koiran, Cosnard & Garzon, 1994). This result can be interpreted and rearranged by saying that Turing machine simulation is possible by piecewise linear systems in dimension three and, hence, is undecidable for such systems. Similar to saturated linear systems, one can use a universal Turing machine to prove that there exists a particular piecewise linear system in R3 (with the state space being partitioned into fewer than 800 subsets) for which is undecidable (Koiran et al., 1994).

Let us now introduce a control variable and consider systems of the form

xt+1=Aixt+But when xtHi.(9)

When B=0, these systems are equivalent to autonomous piecewise linear systems and so is undecidable. This result is also obtained in Blondel and Tsitsiklis (1999a), with a different proof technique that has the advantage that it can be used to derive sharper estimates of the decidability limit (see also Toker, 1996 for a similar proof). There is an obvious tradeoff in piecewise linear systems between the state-space dimension n and the number of subsets m. When there is only one subset, or when the state dimension is equal to one, most properties are easy to check. In Blondel and Tsitsiklis (1999a, Theorem 2), it is shown that is undecidable for piecewise linear systems of dimension 6np+7 on two regions, and that is undecidable for piecewise linear systems of dimension 3np+1 on two regions. Here, np is a number of pairs for which Post's correspondence problem is undecidable and, as explained in Section 2, we can take np=7. This result can be improved by showing that, when nm2+6np, then the problem of deciding whether a given initial state x0 can be driven to the origin, is undecidable. Thus, the problem of interest is undecidable when nm44. In particular, the problem remains undecidable for piecewise linear systems of state dimension 22 and with as few as two subsets.

Finally, the problem , as well as a number of other properties related to stability, are undecidable for piecewise affine systems in dimension two (Blondel, Bournez, Koiran, Papadimitriou & Tsitsiklis, 1999a). The proof makes use of two-counter machines, a model of computation equivalent to Turing machines, as far as decidability is concerned, and relies on the fact that the problem of determining whether a given two-counter machine halts for every initial configuration of the machine is undecidable. If the mapping xt+1=f (xt) is piecewise affine but continuous, the problems related to stability are decidable in dimension 1, undecidable in dimension 3 (Blondel et al., 1999b), and open in dimension 2.

4.5. Other classes of systems

Piecewise linear systems are closely related to the continuous-time piecewise constant derivative systems analyzed in Asarin, Maler and Pnueli (1995). Once more, we are given a partition of Rn into finitely many disjoint subsets H1,...,Hm, and we consider a system of the form

where the bi are given vectors. Assuming that the sets Hi are described by linear inequalities, the trajectories of such systems are piecewise affine functions of time, with break points occurring on the boundaries of the regions. In Asarin et al. (1995) the authors show that, for given states xb and xe, the problem of deciding whether xe is reached by a trajectory starting from xb, is decidable for systems of dimension two, but is undecidable for systems of dimension three or more. Once more, this undecidability result is obtained by simulating Turing machines and by reducing the halting problem to a reachability problem.

Many of the undecidability results presented in this section are based on simulations of Turing machines using dynamical systems of a particular type. Similar Turing machine simulations are possible by other types of devices; see, e.g., Pollack (1987) for simulation by high-order neural nets, 13 Bournez and Cosnard (1996) for simulation by analog automata, Garzon (1995) for simulation by cellular automata, Siegelmann and Sontag (1994) and Siegelmann (1998) for simulation by saturated linear systems that can involve arbitrary real (not necessarily rational) numbers, Branicky (1995b), Branicky (1995a), Ruohonen (1997) and Moore (1991) for simulation by differential equations, Moore (1990) for simulation by continuous-time physical devices in dimension three, and Koiran and Moore (1999) for simulation by analytic maps. In all of these constructions, the state of the system is used to encode the configuration (state and tape contents) of a Turing machine, and the dynamics of the system are used to represent the transition rules of the Turing machine.

The results discussed in this section are far from being exhaustive, and there are several topics that we have not touched. These include systems with polynomial nonlinearities (Sontag, 1988b), bilinear systems ( Sontag, 1988a; Kawski, 1990), decidability results for hybrid systems ( Henzinger, Kopke, Puri & Varaiya, 1998; Lafferierre et al., 1999), and systems evolving on finite groups ( Golaszewski & Ramadge, 1989).

5. Stochastic control

We now turn our attention to problems of stochastic optimal control, and discuss exclusively the discrete-time case. Stochastic control problems can be addressed, in principle, using the methods of dynamic programming. However, their practical applicability is somewhat restricted because many natural problems lead to very large state spaces, a phenomenon that Bellman has termed the "curse of dimensionality." By approaching the subject from the point of view of computational complexity, we can get a lot of insight into the types of problems and formulations that can be efficiently solved. We will mainly focus on Markov decision problems (MDPs), that involve finite state and control spaces, but we will also discuss briefly the case of continuous state spaces.

5.1. Perfectly observed MDPs

A Markov Decision Problem is specified by a finite state space S={1,...,n}, a finite control space U={1,...,m}, a (possibly infinite) time horizon T, a discount factor [0,1], transition probabilities pij(u,t), and one-stage costs g(i,u,t). At a typical time t, the state is equal to some i, a decision u is applied, and the cost g(i,u,t) is incurred. The next state is chosen at random and is equal to j with probability pij(u,t). We will pay special attention to stationary problems, in which the transition probabilities and costs per stage do not depend explicitly on time and can be written as pij(u) and g(i,u).

We define a policy as a sequence of functions t : SU that specify the decision ut at time t as a function of the current state it, that is, ut=t(it). If t is the same for all t, the policy is called stationary, and we abuse terminology by referring to the corresponding function as a policy. Once a policy is fixed, the state of the system evolves as a Markov chain and we define the corresponding expected discounted cost-to-go, as a function of the initial time and state, by

where E[·] stands for expectation with respect to the transition probabilities specified by the policy. When the time horizon T is infinite, we will always assume that the problem is stationary, and that <1, so that the infinite sum converges and the expectation is well-defined. We define the optimal cost-to-go Jt* by

According to the standard theory of dynamic programming, there exists a policy which is optimal, that is, Jt(i)=Jt*(i), for every state i. Furthermore, for infinite-horizon stationary problems, there exists an optimal policy which is stationary, and Jt* is the same for all t. If Jt* can be somehow computed, an optimal policy becomes immediately available by letting

We will mostly refer to complexity results for decision problems where we wish to determine whether Jt*(i) is less than some given rational number, for some given i and t. Typically, the problem of computing an optimal decision t(i) for some given i and t has similar complexity.

For finite-horizon problems, Jt* can be computed by the standard backwards dynamic programming recursion

which is initialized with JT*(i)=0. This algorithm involves O(Tmn2) arithmetic operations, which is of the same order of magnitude as the length of the input (the specification of the transition probabilities pij(u,t) for every i,j,u,t). An interesting twist arises if we deal with stationary problems, because the input now consists of only mn2 transition probabilities and the time horizon T. Since it takes O(log T) bits to specify T, dynamic programming becomes an exponential time algorithm. Still, (Tseng, 1990) shows that for any fixed <1, the problem can be solved with a number of arithmetic operations proportional to log T, provided that the corresponding infinite horizon problem admits a unique optimal policy. (The complexity, however, increases as approaches 1.)

For infinite-horizon, discounted, stationary problems, a solution boils down to solving the Bellman equation

(10)

which is a nonlinear system of n equations, in n unknowns. It is known that the (unique) solution to this equation can be obtained by solving an equivalent linear programming problem (Puterman, 1994; Bertsekas, 1995). Since linear programming can be solved in polynomial time, the same conclusion follows for such MDPs as well. Note, however, that the complexity of an exact solution will generally depend on the numerical values of the input data. No algorithm (neither for linear programming nor for MDPs) is known in which the number of arithmetic operations is a polynomial in n and m, and this is connected to major open problems in linear programming theory (Bertsimas & Tsitsiklis, 1997). This entire discussion applies equally to the average cost MDP, which is a related problem formulation that we will not discuss further. Suffice to say that for both discounted and average cost problems, linear programming is the only method known to solve them in polynomial time.

In practice, MDPs are often solved by special purpose methods, rather than linear programming. The simplest method, value iteration consists of iterating the Bellman equation (10). After k iterations, it is guaranteed to approximate J* to within O(k), and also provides a policy which is within O(k) from being optimal. Because there is a finite number of policies, a policy which is -optimal, for sufficiently small , is guaranteed to be exactly optimal. Building on these observations and certain bounds derived in Tseng (1990), it is shown in Littman et al. (1995) that value iteration can be adapted to solve infinite-horizon discounted MDPs to exact optimality, with computational effort which is polynomial in m, n, 1/(1-), and the size (bit length) of the inputs g(i,u), pij(u). Thus, value iteration is a polynomial time algorithm as long as is fixed to a constant value less than 1.

Some of the fastest methods for infinite-horizon MDPs are based on policy iteration and its refinements. Policy iteration is a method that produces policies with strict performance improvement at each step, until the algorithm terminates (which must eventually happen because there are only finitely many policies). Policy iteration is known to require no more iterations than value iteration, and the complexity analysis in the preceding paragraph still applies. This analysis still allows for the possibility that the number of iterations (policy improvements) increases as approaches 1, but in practice the number of iterations seems to be fairly insensitive to the discount factor. We have just identified a major problem in this area; namely, to determine whether the number of iterations in policy iteration can be bounded by a polynomial in the instance size, or even better, whether the number of iterations is bounded by a polynomial in m and n. The answer to these questions is not known. The authors of Melekopoglou and Condon (1994) consider a variant of policy iteration and show, by means of an example, that the number of policy improvements is in the worst-case exponential in n.14 However, this variant is rather weak in that it can be slower than value iteration, and need not have any implications for the performance of the standard version of policy iteration. Policy iteration bears close resemblance to Newton's method (Puterman & Brumelle, 1978) which, for smooth nonlinear systems of equations, is rapidly convergent in a local sense. Still, this observation is not enough to settle the complexity issue, first because Bellman's equations are not smooth (due to the minimization involved) and, second, because we are interested in convergence in a global sense.

The situation is different when we consider the special case of deterministic problems. In the finite-horizon case, deterministic MDPs are essentially equivalent to the much studied and efficiently solvable shortest path problem. For the infinite-horizon discounted case, the problem can be solved in polynomial time, with a number of arithmetic operations that depends on n but not on the discount factor (Papadimitriou & Tsitsiklis, 1987; Littman, 1996). For the infinite-horizon average cost case, the problem amounts to finding a cycle in a graph that has the smallest possible average arc cost. This is again a classical problem, that can be solved with O(n3) arithmetic operations (Karp, 1978).

5.2. Succintly described problems

Even though MDPs can be solved in time which increases polynomially in the number of states, many problems of practical interest involve a very large number of states, while the problem data (e.g., the transition probabilities) are succinctly described, in terms of a small number of parameters. For concreteness, let us consider the infinite-horizon discounted multi-armed bandit problem. We start with a Markov chain on an n-element state space S, with transition probabilities pij and costs per stage g(i). In a problem involving N arms, the state is an N-tuple (x1,...,xN), where each xk is an element of S and describes the state of the kth arm. The control set is U={1,...,N}. If a decision u=k is made, the state of the kth arm makes a transition to a new state, according to the transition probabilities pij, while the states of the remaining arms stay the same, and a cost g(xk) is incurred. We now have an MDP defined on a state space of cardinality nN, but which is described using only O(n2) parameters (transition probabilities and costs). A straightforward solution using general purpose MDP tools requires at least nN resources (time and memory). However, due to the ingenious solution by Gittins (Gittins, 1989; Bertsekas, 1995), there is an algorithm which given the current state performs a polynomial (in the instance size) amount of computation and determines an optimal action. (Note that there is no point in trying to precompute an optimal action for every state because there are exponentially many states.)

Unfortunately, the multi-armed bandit problem is a rare exception and for most problems of practical interest no alternatives have been found that would allow for computations that are polynomial in the size of a succinct problem description. In fact, a mild variation of the multi-armed bandit problem, the so-called restless bandit problem (Whittle, 1988), becomes PSPACE-hard, even if the problem is deterministic ( Papadimitriou & Tsitsiklis, 1994). For a stronger result, the same reference considers a problem of routing and scheduling in closed queueing networks. An instance of this problem involving n customer classes and servers can be described in terms of O(n2) parameters (service rates, routing probabilities, etc.), but the state space in an MDP formulation is exponentially large. (If each one of n queues can be independently empty or nonempty, we already have 2n states.) It is shown that the problem is EXP-complete, which implies that any algorithm that attempts a short cut ¯¯ similar to the one for the multi-armed bandit problem ¯¯ has provably exponential complexity. This result is developed for average cost problems, but the same proof also works for the finite-horizon case, as long as the time horizon is large (exponential in n). We also refer to Mundhenk et al. (1997, Theorem 6.7) and Littman (1997, Theorem 1) which study succinctly described exponential or infinite horizon MDPs, and provide EXP-completeness results consistent with the one discussed here. The interest on this subject partly stems from recent research activity on propositional planning systems within the field of artificial intelligence. It is shown in Littman (1997) that the EXP-completeness result applies to several classes of planning systems in the literature. Finally, note that if we restrict attention to deterministic succinctly described MDPs, the queueing network problem becomes PSPACE-complete ( Papadimitriou and Tsitsiklis, 1994, Theorem 3), which is consistent with results on planning problems in deterministic domains (Bylander, 1994; Littman, Goldsmith & Mundhenk, 1998).

Another case of interest refers to succinctly described MDPs over a time horizon which is polynomial in the instance size, and therefore logarithmic in the number of states. This is motivated in the planning literature by the fact that excessively long plans cannot be practically useful. For this case, the problem becomes PSPACE-complete. This is proved in Littman (1997, Theorems 2 and 3) for several planning systems, and in Mundhenk et al. (1997, Corollary 6.14) for a related class of "compressed" MDPs; see also Littman et al. (1998).

5.3. Problems with continuous state spaces

Let us briefly discuss some related work on problems with continuous state spaces. Problems of this type do not admit closed-form or exact algorithmic solutions. A rare exception is the problem of optimal control of linear systems, when the cost per stage is a nonnegative quadratic function of the state and the control, and this is because the optimal cost-to-go function turns out to be quadratic. The finite horizon problem can be solved exactly using the standard dynamic programming recursion. The infinite horizon problem amounts to solving an algebraic Riccati equation, which cannot be done exactly, but there are rapidly convergent algorithms that can quickly produce solutions within any desired precision.

General continuous-state problems can be solved approximately, by discretizing them, as long as the problem data (transition probabilities and cost per stage) are sufficiently smooth functions of the state (Whitt and Whitt). References Chow and Tsitsiklis (1989) and Chow and Tsitsiklis (1991) establish that

arithmetic operations are necessary and sufficient for uniformly approximating the function J* within , for the case where the state and control spaces are the sets [0,1]n and [0,1]m, respectively, under a Lipschitz continuity assumption on the problem data, and for a fixed discount factor .15 The 1/m complexity arises because we need to search a grid with spacing in order to find a close-to-optimal decision for any given state. This issue disappears if the control set has finite cardinality. However, the curse of dimensionality remains in effect due to the 1/2n term.

In some important work, Rust (1997a) has shown that the curse of dimensionality can be bypassed by allowing for randomized algorithms. The advantage of randomization can be best understood in terms of the multivariable integration problem. Deterministic methods require a fine discretization of the space of interest (which leads to the curse of dimensionality), whereas Monte Carlo methods produce -approximate estimates of the value of an integral, with high probability, using a dimension-independent number O(1/2) of samples. Rust's method is a variant of value iteration which only looks at a relatively small number of randomly sampled representative states.

The complexity estimates in all of these references are proportional to the Lipschitz constant of the problem data. When one considers practical problems of increasing dimension, the Lipschitz constant will often increase exponentially. To see this, consider the probability density function p(x)=2x for a scalar random variable that takes values in [0,1]. If we introduce n independent copies of this random variable, the density becomes 2nx1···xn on the unit cube and the Lipschitz constant increases exponentially with n. Thus, the curse of dimensionality is not fully eliminated. Nevertheless, the results of Rust suggest a definite advantage in using random sampling (or even deterministic but pseudo-random sampling (Rust, 1997b)).

5.4. Problems with imperfect information

We now return to finite-state systems. So far, we have allowed the decision u to be a function of the system state. In an alternative formulation, the controller only has access to partial observations yt=h(it), where h is a prespecified function on the state space, which we assume to have finite range. (One could augment this model by including observation noise. However, by suitably redefining the state of the system, the problem with observation noise can be reduced to a problem without it.) In the standard formulation of the partially observed Markov decision problem (POMDP), the decision ut at time t is allowed to be a function of all past observations; that is, we are dealing with policies of the form ut=t(y0,y1,...,yt).

POMDPs can be reduced to perfectly observed MDPs, by redefining the state at time t to be the past observation history (y0,y1,...,yt) (Bertsekas, 1995). With a time horizon T, the cardinality of the redefined state space (and the resulting dynamic programming algorithm) increases exponentially in T, but this is to some extent unavoidable. Indeed, if we assume that the original state space has cardinality n and let T=n, the problem is PSPACE-complete (Papadimitriou & Tsitsiklis, 1987, Theorem 6). Incidentally, the MDP obtained in this way is another example of a succinctly described MDP, with a time horizon which is polynomial in the size of the instance, and so this result is in agreement with the results mentioned in Section 5.2.

Another standard reformulation of an n-state POMDP to a perfectly observed MDP uses a redefined state (the "information state") which is the posterior distribution of the original state, given the available observations. This is now an MDP with a continuous state space (the unit simplex in Rn), and smooth transition probabilities, and it can be solved approximately (cf. the discussion in Section 5.3). If the preimage of every possible observation has cardinality bounded by a constant k, and if k3, the problem remains NP-hard. (The case k=2 is open.) But we are now dealing with an essentially k-dimensional state space and for any fixed k, the problem can be solved to within of optimality with a number of arithmetic operations which is polynomial in n, m, 1/, and the time horizon (Burago, Rougement & Slissenko, 1996, Theorem 8).

Infinite-horizon POMDPs cannot be solved exactly by reducing them to perfectly observed MDPs, because the latter have an infinite state space. It turns out that infinite-horizon POMDPs are undecidable (under total, discounted, or average cost criteria) (Madani, Hanks & Condon, 1999), because they are closely related to the "emptiness" problem for probabilistic finite-state automata, which is known to be undecidable. There are also results for succinctly represented finite-horizon POMDPs. Generally speaking, the complexity (as a function of the instance size) is exponentially higher ( Mundhenk et al., 1997).

Let us also mention an interesting special case of POMDPs, namely, the case where there are no observations (open-loop control). Assuming a time horizon T equal to the cardinality of the state space, the problem is NP-complete (Papadimitriou and Tsitsiklis, 1987, Corollary 2). See also Mundhenk et al. (1997) for results concerning the case of succinct problem descriptions, and Burago et al. (1996, Theorem 6) for nonapproximability results.

At this point, it is interesting to draw a connection with some of the results discussed in Section 3. For any decision u, let P(u) be the corresponding transition probability matrix. Let be a row vector with the probability distribution of the initial state, and let g be a column vector whose ith component is the cost if the system is found at state i at time T. (Assume that there are no costs at intermediate times.) Optimal open-loop control amounts to minimizing P(u0)P(u1)···P(uT-1)g over all sequences u0,u1,...,uT-1, and as stated earlier, this is an NP-complete problem. But this problem can also be viewed as referring to the worst-case behavior of a time-varying linear system xt+1=Ptxt, where each matrix Pt is chosen from within a finite set of matrices. Indeed, this observation has been exploited in Tsitsiklis and Blondel (1997a) to establish NP-hardness results for problems related to time-varying linear systems (cf. Section 3.5).

5.5. Nonclassical information structures

In stochastic control problems with a nonclassical information structure, the decision at time t is made without knowledge of the full history of past observations, and this setting includes problems of decentralized stochastic control and team decision theory. When arbitrary information structures are allowed, POMDPs are obtained as a special case and therefore, the complexity is at least as large. But negative results can be obtained even for seemingly simple special cases.

Consider the static team decision problem (Radner, 1962). Decision maker 1 (respectively, 2) observes a random variable y1 (respectively, 2) and makes a decision according to a rule of the form u1=1(y1) (respectively, u2=2(y2)), where u1 and u2 are restricted to lie in finite sets. We consider the objective

and ask the question whether it is less than a given rational number. Here, y1 and y2 are finite-valued random variables with a given (rational-valued) probability mass function, and g is a given (rational-valued) cost function. This problem was shown NP-complete in Papadimitriou and Tsitsiklis (1982), and this result remains valid for a special case that arises in the context of decentralized detection ( Tsitsiklis, 1984; Tsitsiklis and Athans, 1985).

The same problem can be recast as an MDP with a nonclassical information structure, and a time horizon T=3. We view (y1,y2) as the state after the first transition. At stages 1 and 2, y1 and y2 are observed, respectively, but we require the decisions at any time t to be functions of the present observation only, of the form ut=t(yt), t=1,2. Thus, imperfectly observed MDPs with a nonclassical information structure are NP-complete even when the time horizon is T=3. (With a time horizon of two, and a fixed initial state, we are back to the case of a classical information structure and such problems are easily seen to be polynomial-time solvable.)

Problems in which we restrict to policies of the form ut=t(yt) are referred to as "time-dependent partially observable" in Mundhenk et al. (1997). This reference characterizes the complexity of several related variants; see also Littman (1994). We also refer the reader to Papadimitriou and Tsitsiklis (1986) which establishes the NP-completeness of a discrete variant of Witsenhausen's counterexample ( Witsenhausen, 1968). 16

5.6. Supervisory control of discrete-event systems

There is an extensive literature on supervisory control of discrete-event systems, within the framework of Ramadge (1983) and Ramadge and Wonham (1987). Many of the problem formulations in this area lead to problems similar to MDPs, except that the transitions are not determined stochastically, but may be chosen by an adversary. The difference between the two model types is rather minor, because a variant of dynamic programming, applicable to min¯max problems, can be used. In particular, the standard formulation of the supervisory control problem can be solved in polynomial time if the supervisor has perfect information ( Ramadge and Wonham, 1987), but becomes PSPACE-hard for the case of imperfect information, because it has essentially the same structure as POMDPs ( Tsitsiklis, 1989). See also Rudie and Willems (1995) for some related results.

5.7. Zero-sum Markov games

The setting here is similar to MDPs except that there are two competing decision makers, one acting at odd times, the other at even times. One decision maker wants to maximize the total expected cost, and the other attempts to minimize it. The finite-horizon version of the problem is similar to MDPs: with perfect information, it can be solved by a simple adaptation of the dynamic programming recursion. For the infinite-horizon discounted case, the basic theory, including a suitable Bellman equation is again the same as for MDPs (Shapley, 1953; Raghavan and Filar, 1991). However, a reformulation based on linear programming is not known, and unlike ordinary MDPs, a polynomial-time algorithm is not available. On the other hand, a negative complexity result is not available either, and this problem's complexity is an important open problem. We also refer the reader to Condon and Condon and Zwick and Paterson (1996) for complexity results on more general classes of games.

6. Conclusions

The subject of complexity in systems and control is multifaceted and interesting in many different ways. It can help the practitioner in choosing problem formulations, and in calibrating expectations of what can be algorithmically accomplished. For the systems theorist or the applied mathematician, it raises a variety of challenging open problems that require a diverse set of tools from both discrete and continuous mathematics. Finally, for the theoretical computer scientist, the problems in systems and control theory provide the opportunity to relate abstractly defined complexity classes with specific problems of practical interest.

Acknowledgements

We gratefully acknowledge the input of several people, including Eugen Asarin, Olivier Bournez, Michael Branicky, Laurent El Ghaoui, Minyue Fu, Stéphane Gaubert, Leonid Gurvits, Pascal Koiran, Michael Littman, Oded Maler, Maurice Mignotte, Cristopher Moore, Jiri Rohn, John Rust, Rodolphe Sepulchre, Eduardo Sontag, Roberto Tempo, Jan van Schuppen, M. Vidyasagar and Louis Wehenkel. Also, the second author has greatly benefitted from collaborating with Christos Papadimitriou over the years.

References

Aho, A.V., Hopcroft, J.E. and Ullman, J.D., 1974. The design and analysis of computer algorithms, Addison-Wesley, Reading, MA.

Alur, R., Courcoubetis, C., Halbwachs, N., Henzinger, T.A., Ho, P.-H., Nicollin, X., Olivero, A., Sifakis, J. and Yovine, S., 1995. The algorithmic analysis of hybrid systems. Theoretical Computer Science 138, pp. 3¯34. Abstract | Journal Format-PDF (1164 K) | MathSciNet

Anderson, B.D.O., Bose, N.K. and Jury, E.I., 1975. Output feedback stabilization and related problem ¯¯ solutions via decision methods. IEEE Transactions on Automatic Control 20, pp. 53¯66. INSPEC Compendex  | MathSciNet

Asarin, E., Maler, O. and Pnueli, A., 1995. Reachability analysis of dynamical systems having piecewise-constant derivatives. Theoretical Computer Science 138, pp. 35¯66.

Baccelli, F., Cohen, G., Olsder, G.-J. and Quadrat, J.-P., 1992. Synchronization and linearity, Wiley, New York.

Barabanov, N. E. (1988). Lyapunov indicators of discrete inclusions, parts I, II and III, Avtomatika i Telemekhanika, 2, 40¯46, 3, 24¯29 and 5, 17¯24. Translation in Automatic remote control (1988) part I, 49(3), 152¯157, part II, 49(5), 558¯565.

Barmish, B. R., Floudas, C. A., Hollot, C. V., & Tempo, R. (1994). A global linear programming solution to some open robustness problems. Proceedings of the American control conference, Seattle, USA (pp. 3871¯3877).

Bartlett, A.C., Hollot, C.V. and Huang, L., 1988. Root locations of an entire polytope of polynomials: It suffices to check the edges. Mathematics of Control, Signals, and Systems 1, pp. 61¯71. MathSciNet

Basu, S., Pollack, R. and Roy, M.-F., 1996. On the combinatorial and algebraic complexity of quantifier elimination. Journal of the ACM 43, pp. 1002¯1046.

Berger, M. and Wang, Y., 1992. Bounded semigroups of matrices. Linear Algebra Applications 166, pp. 21¯27. MathSciNet

Bertsekas, D.P., 1995. Dynamic programming and optimal control, Athena Scientific, Belmont, MA.

Bertsimas, D. and Tsitsiklis, J.N., 1997. Introduction to linear optimization, Athena Scientific, Belmont, MA.

Bialas, S., 1985. A necessary and sufficient condition for the stability of convex combinations of stable polynomials or matrices. Bulletin of the Polish Academy of Sciences, Technical Sciences 33, pp. 473¯480. Compendex  | MathSciNet

Blondel, V. D., Bournez, O., Koiran, P., Papadimitriou, C., & Tsitsiklis, J. N. (1999a). Deciding stability and mortality of piecewise affine dynamical systems. Preprint.

Blondel, V. D., Bournez, O., Koiran, P., & Tsitsiklis, J. N. (1999b). The stability of saturated linear dynamical systems is undecidable. Preprint.

Blondel, V. D., Gaubert, S., & Tsitsiklis, J. N. (1998). Approximating Lyapunov exponents of matrices in the max-algebra is NP-hard. Preprint.

Blondel, V.D., Gevers, M. and Lindquist, A., 1995. Survey on the state of systems and control. European Journal of Control 1, pp. 5¯23. Compendex INSPEC 

Blondel, V. D., & Portier, N. (1999a). The presence of a zero in an integer linear recurrent sequence is NP-hard to decide. Preprint.

Blondel, V.D. and Tsitsiklis, J.N., 1997. NP-hardness of some linear control design problems. SIAM Journal of Control and Optimization 35, pp. 2118¯2127. Compendex INSPEC  | MathSciNet

Blondel, V.D. and Tsitsiklis, J.N., 1997. When is a pair of matrices mortal?. Information Processing Letters 63, pp. 283¯286. Abstract | Journal Format-PDF (240 K) | MathSciNet

Blondel, V. D., & Tsitsiklis, J. N. (1998). Overview of complexity and decidability results for three classes of elementary nonlinear systems. In Y. Yamamoto, & S. Hara, Learning, control and hybrid systems (pp. 46¯58) Heidelberg: Springer.

Blondel, V.D. and Tsitsiklis, J.N., 1999. Complexity of stability and controllability of elementary hybrid systems. Automatica 35, pp. 479¯489. Abstract | Journal Format-PDF (185 K)

Blondel, V. D., & Tsitsiklis, J. N. (1999b). Boundedness of finitely generated matrix semigroups is undecidable. Preprint.

Blondel, V. D., Sontag, E. D., Vidyasagar, M., & Willems, J. C. (1999c). Open problems in mathematical systems and control theory. London: Springer, http://www.inma.ucl.ac.be/~blondel/openprobs.html.

Blum, L., Cucker, F., Shub, M. and Smale, S., 1998. Complexity and real computation, Springer, New York.

Bournez, O. and Cosnard, M., 1996. On the computational power of dynamical systems and hybrid systems. Theoretical Computer Science 168, pp. 417¯459. Abstract | Journal Format-PDF (1908 K) | MathSciNet

Bournez, O., & Branicky, M. (1998). On matrix mortality in low dimensions. Technical Report 98-01, Verimag, France.

Boyd, S., El Ghaoui, L., Feron, E., & Balakrishnan, V. (1994). Linear matrix inequalities in system and control theory. SIAM Studies in Applied Mathematics, vol. 15. Philadelphia, PA: SIAM.

Braatz, R., Young, P., Doyle, J. and Morari, M., 1994. Computational complexity of calculation. IEEE Transactions on Automatic Control 39, pp. 1000¯1002. Abstract  | MathSciNet

Branicky, M.S. (1995a). Studies in hybrid systems: Modeling, analysis, and control. Ph.D. thesis, Center for Intelligent Control Systems, Cambridge, MA: MIT Press.

Branicky, M.S., 1995. Universal computation and other capabilities of hybrid and continuous dynamical systems. Theoretical Computer Science 138, pp. 67¯100. Abstract | Journal Format-PDF (1332 K) | MathSciNet

Brayton, R. and Tong, C., 1980. Constructive stability and asymptotic stability of dynamical systems. IEEE Transactions on Circuits and Systems 27, pp. 1121¯1130. INSPEC Compendex  | MathSciNet

Bylander, T., 1994. The computational complexity of propositional STRIPS planning. Artificial Intelligence 69, pp. 165¯204. Abstract  | MathSciNet

Chow, C.-S. and Tsitsiklis, J.N., 1989. The complexity of dynamic programming. Journal of Complexity 5, pp. 466¯488. MathSciNet

Chow, C.-S. and Tsitsiklis, J.N., 1991. An optimal one-way multigrid algorithm for discrete-time stochastic control. IEEE Transactions on Automatic Control 36, pp. 898¯914. INSPEC Compendex 

Condon, A., 1989. Computational models of games, The MIT Press, Cambridge, MA.

Condon, A., 1992. The complexity of stochastic games. Information and Computation 96, pp. 203¯224. INSPEC Compendex  | MathSciNet

Cook, S.A. (1971). The complexity of theorem proving procedures. Proceedings of the third ACM symposium on the theory of computing (pp. 117¯128).

Coxson, G.E. and De Marco, C.L., 1994. The computational complexity of approximating the minimal perturbation scaling to achieve instability in an interval matrix. Mathematics of Control, Signals, and Systems 7, pp. 279¯291. INSPEC INSPEC  | MathSciNet

Coxson, G. E. (1993). Computational complexity of robust stability and regularity in families of linear systems. Ph.D. thesis, Electrical and Computer Engineering Department, University of Wisconsin at Madison.

Dasgupta, S., 1988. Kharitonov's theorem revisited. Systems and Control Letters 11, pp. 381¯384. INSPEC INSPEC  | MathSciNet

Daubechies, I. and Lagarias, J.C., 1992. Sets of matrices all infinite products of which converge. Linear Algebra Applications 162, pp. 227¯263. MathSciNet

Davis, M., 1973. Hilbert's tenth problem is unsolvable. American Mathematical Monthly 80, pp. 233¯269. MathSciNet

Davis, M., 1982. Computability and unsolvability, Dover, New York.

Demmel, J.W., 1992. The component-wise distance to the nearest singular matrix. SIAM Journal of Matrix Analysis and Applications 13, pp. 10¯19. MathSciNet

Doyle, J.C., 1982. Analysis of feedback systems with structured uncertainties. IEE Proceedings 129, pp. 240¯250.

Fu, M. (1999). Approximation of complex . In V. D. Blondel, E. D. Sontag, M. Vidyasagar, & J. C. Willems, Open problems in mathematical systems and control theory, Problem 22. London: Springer.

Fu, M. and Luo, Z.-Q., 1997. Computational complexity of a poblem aising in fixed order output feedback design. Systems and Control Letters 30, pp. 209¯215. MathSciNet

Fu, M. and Barmish, B.R., 1988. Maximal unidirectional perturbation bounds for stability of polynomials and matrices. Systems and Control Letters 11, pp. 173¯179. MathSciNet

Fu, M., & Dasgupta, S. (1998). Computational complexity of real structured singular value in p setting. Preprint.

Fu, M., Dasgupta, S. and Blondel, V., 1995. Robust stability under a class of nonlinear parametric perturbations. IEEE Transactions on Automatic Control 40, pp. 213¯223. Abstract  | MathSciNet

Gantmacher, F.R. (1959). Matrix theory, vol. 2. New York: Chelsea Publishing Company.

Garey, M.R. and Johnson, D.S., 1979. Computers and intractability: A guide to the theory of NP-completeness, Freeman and Co, New York.

Garzon, M., 1995. Models of massive parallelism. Analysis of cellular automata and neural networks, Springer, Berlin.

Gittins, J.C., 1989. Multi-armed bandit allocation indices, Wiley, New York.

Golaszewski, C. and Ramadge, P.J., 1989. The complexity of some reachability problems for a system on a finite group. Systems and Control Letters 12, pp. 431¯435. INSPEC Compendex  | MathSciNet

Gripenberg, G., 1996. Computing the joint spectral radius. Linear Algebra Applications 234, pp. 43¯60. Abstract | Journal Format-PDF (450 K) | MathSciNet

Gurvits, L., 1995. Stability of discrete linear inclusions. Linear Algebra Applications 231, pp. 47¯85. Abstract | Journal Format-PDF (953 K) | MathSciNet

Gurvits, L. (1996). Stability of linear inclusions ¯¯ Part 2. Technical Report TR 96-173, NEC Research.

Henzinger, T., Kopke, P., Puri, A. and Varaiya, P., 1998. What's decidable about hybrid automata. Journal of Computer and System Science 57, pp. 97¯124.

Hertz, D., 1992. The extreme eigenvalues and stability of real symmetric interval matrices. IEEE Transactions on Automatic Control 37, pp. 532¯535. INSPEC Compendex  | MathSciNet

Hooper, P., 1966. The undecidability of the Turing machine immortality problem. The Journal of Symbolic Logic 2, pp. 219¯234. MathSciNet

Hopcroft, J.E. and Ullman, J.D., 1969. Formal languages and their relation to automata, Addison-Wesley, Reading, MA.

Hopcroft, J.E. and Ullman, J.D., 1979. Introduction to automata theory, languages, and computation, Addison-Wesley, Reading, MA.

Hyotyniemu, H. (1997). On unsolvability of nonlinear system stability. Proceedings of the fourth European control conference, Brussels, vol. 636 (pp. 1¯6).

Jones, J.P., 1974. Recursive undecidability. American Mathematical Monthly 81, pp. 724¯738. MathSciNet

Karp, R. M. (1972). Reducibility among combinatorial problems. In R. Miller, & J. Thatcher, Complexity of computer computations. New York: Plenum Press (pp. 85¯104).

Karp, R.M., 1978. A characterization of the minimum cycle mean in a digraph. Discrete Mathematics 23, pp. 309¯311. INSPEC  | MathSciNet

Kawski, M., 1990. The complexity of deciding controllability. Systems and Control Letters 15, pp. 9¯14. INSPEC Compendex  | MathSciNet

Khargonekar, P. P., & Tikku, A. (1996). Randomized algorithms for robust control analysis have polynomial complexity. Proceedings of the conference on decision and control.

Kharitonov, V.L., 1978. Asymptotic stability of an equilibrium position of a family of systems of linear differential equations. Differentsial'nye Uravneniya 14, pp. 2086¯2088.

Kilian, J. and Siegelmann, H., 1996. The dynamic universality of sigmoidal neural networks. Information and Computation 128, pp. 48¯56. INSPEC  | MathSciNet | CrossRef

Klarner, D.A., Birget, J.-C. and Satterfield, W., 1991. On the undecidability of the freeness of integer matrix semigroups. International Journal of Algebra and Computation 1, pp. 223¯226. INSPEC  | MathSciNet

Koiran, P., 1996. A family of universal recurrent networks. Theoretical Computer Science 168, pp. 473¯480. Abstract | Journal Format-PDF (316 K) | MathSciNet

Koiran, P. and Moore, C., 1999. Closed-form analytic maps in 1 or 2 dimension can simulate universal Turing machines. Theoretical Computer Science 210, pp. 217¯223. Abstract | Journal Format-PDF (266 K) | MathSciNet

Koiran, P., Cosnard, M. and Garzon, M., 1994. Computability properties of low-dimensional dynamical systems. Theoretical Computer Science 132, pp. 113¯128. Abstract  | MathSciNet

Kokame, H., & Mori, T. (1992). A branch-and-bound method to check the stability of a polytope of matrices. In M. Mansour, S. Balemi & W. Truoel, Robustness of dynamics systems with parameter uncertainties (pp. 125¯137). Basel: Birkhauser.

Kozyakin, V.S., 1990. Algebraic unsolvability of problem of absolute stability of desynchronized systems. Automation and Remote Control 51, pp. 754¯759. Compendex  | MathSciNet

Kreinovich, V., Lakeyev, A., & Rohn, J. (1995). Computational complexity of interval algebraic problems: Some are feasible and some are computationally intractable ¯¯ a survey. Scientific computing and validated numerics (Wuppertal, 1995) (pp. 293¯306). Mathematical Research, vol. 90. Berlin: Akademie Verlag.

Kreinovich, V., Lakeyev, A., Rohn, J., & Kahl, P. (1997). Computational complexity and feasibility of data processing and interval computation. Applied Optimization Series, vol. 10. Dordrecht: Kluwer Academic.

Krom, M., 1981. An unsolvable problem with products of matrices. Mathematical Systems Theory 14, pp. 335¯337. INSPEC  | MathSciNet

Krom, M. and Krom, M., 1989. Recursive solvability of problems with matrices. Z. Math. Logik Grundlagen Math. 35, pp. 437¯442. INSPEC  | MathSciNet

Lafferierre, G., Pappas, G. J., & Yovine, S. (1999). A new class of decidable hybrid systems. Lecture Notes in Computer Sciences, vol. 1596. Heidelberg: Spirnger.

Lagarias, J.C. and Wang, Y., 1995. The finiteness conjecture for the generalized spectral radius of a set of matrices. Linear Algebra Applications 214, pp. 17¯42. Abstract | Journal Format-PDF (777 K) | MathSciNet

Littman, M.L., 1994. Memoryless policies: Theoretical limitations and practical results. In: Cliff, D., Husbands, P., Meyer, J.-A. and Wilson, S.W., Editors, 1994. From animals to animats 3: Proceedings of the third International Conference on simulation of adaptive behavior, The MIT Press, Cambridge, MA.

Littman, M. L. (1996). Algorithms for sequential decision making. Ph.D. thesis, Department of Computer Science, Brown University.

Littman, M. L. (1997). Probabilistic propositional planning: Representations and complexity. In Proceedings of the 14th national conference on artificial intelligence (pp. 748¯754). Cambridge, MA: AAAI Press/The MIT Press.

Littman, M. L., Dean, T. L., & Kaelbling, L. P. (1995). On the complexity of solving Markov decision problems. Proceedings of the 11th annual conference on uncertainty in artificial intelligence, Montreal, Québec, Canada (pp. 394¯402).

Littman, M.L., Goldsmith, J. and Mundhenk, M., 1998. The computational complexity of probabilistic plan existence and evaluation. Journal of AI Research 9, pp. 1¯36. INSPEC  | MathSciNet

Liu, D. and Michel, A., 1994. Dynamical systems with saturation nonlinearities: Analysis and design, Springer, London.

Madani, O., Hanks, S., & Condon, A. (1999). On the undecidability of probabilistic planning and infinite-horizon partially observable Markov decision problems, Proceedings of AAAI99, to appear.

Matiyasevich, Yu., 1970. Enumerable sets are diophantine. Doklady Akademii Nauk SSSR 191, pp. 279¯282.

Matiyasevich, Y., & Sénizergues, G. (1996). Decision problem for semi-Thue systems with a few rules. Proceedings of LICS96 (pp. 523¯531). Silver Spring, MD: IEEE Computer Society Press.

Melekopoglou, M. and Condon, A., 1994. On the complexity of the policy improvement algorithm for markov decision processes. ORSA Journal on Computing 6, pp. 188¯192. INSPEC 

Mignotte, M., Shorey, T.N. and Tijdeman, R., 1984. The distance between terms of an algebraic recurrence sequence. Journal fur die Reine and Angewandte Mathematik 349, pp. 63¯76. MathSciNet

Minnichelli, R.J., Anagnost, J.J. and Desoer, C.A., 1989. An elementary proof of Kharitonov's theorem with extensions. IEEE Transactions on Automatic Control 34, pp. 995¯998. INSPEC Compendex  | MathSciNet

Minsky, M., 1967. Computation. Finite and infinite machines, Prentice-Hall, Englewood Cliffs, NJ.

Moore, C., 1990. Unpredictability and undecidability in dynamical systems. Physical Review Letters 64, pp. 2354¯2357. INSPEC  | MathSciNet

Moore, C., 1991. Generalized shifts: Undecidability and unpredictability in dynamical systems. Nonlinearity 4, pp. 199¯230. INSPEC  | MathSciNet

Moore, C. (1998). Finite-dimensional analog computers: Flows, maps, and recurrent neural networks. Proceedings of the first international conference on unconventional models of computation, Auckland, New Zealand.

Mundhenk, M. J., Goldsmith, J., Lusena, C., & Allender, E. (1997). Encyclopedia of complexity results for finite horizon Markov decision process problems. Technical Report 273-97, Computer Science Department, University of Kentucky.

Nemhauser, G.L. and Wolsey, L.A., 1988. Integer and combinatorial optimization, Wiley, New York.

Nemirovskii, A., 1993. Several NP-Hard problems arising in robust stability analysis. Mathematics of Control. Signals, and Systems 63, pp. 99¯105. INSPEC Compendex  | MathSciNet

Nemirovsky, A.S. and Yudin, D.B., 1983. Problem complexity and method efficiency in optimization, Wiley, New York.

Orponen, P., 1994. Computational complexity of neural networks: A survey. Nordic Journal of Computing 1, pp. 94¯100.

Papadimitriou, C.H., 1994. Computational complexity, Addison-Wesley, Reading, MA.

Papadimitriou, C.H. and Tsitsiklis, J.N., 1982. On the complexity of designing distributed protocols. Information and Control 53, pp. 211¯218. INSPEC Compendex  | MathSciNet

Papadimitriou, C.H. and Tsitsiklis, J.N., 1986. Intractable problems in control theory. SIAM Journal of Control and Optimization 24, pp. 639¯654. Compendex  | MathSciNet

Papadimitriou, C.H. and Tsitsiklis, J.N., 1987. The complexity of Markov decision processes. Mathematics of Operations Research 12, pp. 441¯450. INSPEC INSPEC  | MathSciNet

Papadimitriou, C. H., & Tsitsiklis, J. N. (1994). The complexity of optimal queueing network control. Proceedings of the nineth annual conference on struture in complexity theory. Amsterdam, The Netherlands. Full version available from: http://web.mit.edu/jnt/www/publ.html.

Paterson, M., 1970. Unsolvability in 3×3 matrices. Studies in Applied Mathematics 49, pp. 105¯107. MathSciNet

Poljak, S. and Rohn, J., 1993. Checking robust nonsingularity is NP-hard. Mathematics of Control, Signals, and Systems 6, pp. 1¯9. INSPEC Compendex  | MathSciNet

Pollack, J. B. (1987). On connectionist models of natural language processing. Ph.D. thesis, Department of Computer Science, University of Illinois, Urbana.

Post, E.L., 1946. A variant of a recursively unsolvable problem. Bulletin of the American Mathematical Society 52, pp. 264¯268. MathSciNet

Puterman, M.L., 1994. Markov decision processes, Wiley, New York.

Puterman, M. L., & Brumelle, S. L. (1978). The analytic theory of policy iteration. In M.L. Puterman, Dynamic programming and its applications. New York: Academic Press.

Raghavan, T.E.S. and Filar, J.A., 1991. Algorithms for stochastic games ¯¯ a survey. Zeitschrift für O. R. 35, pp. 437¯472. INSPEC  | MathSciNet

Radner, R., 1962. Team decision problems. Annals of Mathematical Statistics 33, pp. 857¯881. MathSciNet

Ramadge, P. J. (1983). Control and supervision of discrete event processes. Ph.D. thesis, Department of Electrical Engineering, University of Toronto.

Ramadge, P.J. and Wonham, W.M., 1987. Supervisory control of a class of discrete event processes. SIAM J. Control Optimization 25, pp. 206¯230. INSPEC Compendex  | MathSciNet

Rogozhin, Y., 1996. Small universal Turing machines. Theoretical Computer Science 168, pp. 215¯240. Abstract | Journal Format-PDF (793 K) | MathSciNet

Rohn, J., 1994. Positive definiteness and stability of interval matrices. SIAM Journal on Matrix Analysis and Applications 15, pp. 175¯184. MathSciNet

Rohn, J., 1994. Checking positive definiteness or stability of symmetric interval matrices is NP-hard. Commentationes Mathematicae Universitatis Carolinae 35, pp. 795¯797. MathSciNet

Rohn, J., 1997. Complexity of some linear problems with interval data. Reliable Computing 3, pp. 315¯323. INSPEC  | MathSciNet

Rosenthal, J., & Wang, X. (1997). Inverse eigenvalue problems for multivariable linear systems. In C. I. Byrnes, B. N. Datta, D. Gilliam, & C. F. Martin, Systems and control in the twenty-first century (pp. 289¯311). Basel: Birkhauser.

Rota, G.-C. and Strang, G., 1960. A note on the joint spectral radius. Nederlandse Koninklÿke Akademie & van Wetenschappen Indagationes Mathematicae 22, pp. 379¯381. MathSciNet

Rudie, K. and Willems, J.C., 1995. The computational complexity of decentralized discrete-event control problems. IEEE Transactions on Automatic Control 40, pp. 1313¯1319. Abstract 

Ruohonen, K., 1997. Undecidable event detection problems for ODEs of dimension one and two. RAIRO Information and Theoretical Applications 31, pp. 67¯79. INSPEC  | MathSciNet

Rust, J. (1996). Numerical dynamic programming in economics. In H. Amman, D. Kendrick, & J. Rust, Handbook of computational economics (pp. 619¯729). Amsterdam: Elsevier (Chapter 14).

Rust, J., 1997. Using randomization to break the curse of dimensionality. Econometrica 65 3, pp. 487¯516. MathSciNet

Rust, J. (1997b). A comparison of policy iteration methods for solving continuous-state, infinite-horizon Markovian decision problems using random, quasi-random, and deterministic discretizations. Available from the Economics Working Paper archive: http://econwpa.wustl.edu/eprints/comp/papers/9704/9704001.abs.

Safonov, M.G., 1982. Stability margins of diagonally perturbed multivariable feedback systems. Proceedings of IEE-D 129, pp. 251¯256. INSPEC Compendex  | MathSciNet

Salomaa, A., 1985. Computation and automata, Cambridge University Press, Cambridge.

Salomaa, A. and Soittola, M., 1978. Automata-theoretic aspects of formal power series, Springer, New York.

Samake, K. (1996). Suite récurrentes linéaires, problèmes d'effectivité. Ph.D. thesis, Institut de Recherche Mathématique Avançée, Université Louis Pasteur, Strasbourg, France.

Schrijver, A., 1986. Theory of linear and integer programming, Wiley, New York.

Seidenberg, A., 1954. A new decision method for elementary algebra. Annals of Mathematics 60, pp. 365¯374. MathSciNet

Shapley, L.S., 1953. Stochastic games. Proceedings of the National Academy of Sciences 39, pp. 1095¯1100. MathSciNet

Sideris, A., 1991. An efficient algorithm for checking the robust stability of a polytope of polynomials. Mathematics of Control, Signals, and Systems 4, pp. 315¯337. INSPEC Compendex  | MathSciNet

Siegelmann, H.T., 1998. Neural networks and analog computation. Beyond the Turing limit, Birkhauser, Boston.

Siegelmann, H.T. and Sontag, E.D., 1991. Turing computability with neural nets. Applied Mathematics Letters 4, pp. 77¯80. MathSciNet

Siegelmann, H. and Sontag, E., 1994. Analog computation via neural networks. Theoretical Computer Science 131, pp. 331¯360. Abstract  | MathSciNet

Siegelmann, H. and Sontag, E., 1995. On the computational power of neural nets. Journal of Computer and System Sciences, pp. 132¯150. Abstract  | MathSciNet | CrossRef

Soh, C.B., 1990. Necessary and sufficient conditions for stability of symmetric interval matrices. International Journal of Control 51, pp. 243¯248. INSPEC Compendex  | MathSciNet

Sontag, E.D., 1981. Nonlinear regulation: the piecewise linear approach. IEEE Transactions on Automatic Control 26, pp. 346¯358. INSPEC Compendex  | MathSciNet

Sontag, E.D., 1985. Real addition and the polynomial hierarchy. Information Processing Letters 20, pp. 115¯120. INSPEC Compendex  | MathSciNet

Sontag, E.D., 1988. Controllability is harder to decide than accessibility. SIAM Journal of Control and Optimization 26, pp. 1106¯1118. INSPEC Compendex  | MathSciNet

Sontag, E. D. (1988b). Some complexity questions regarding controllability. Proceedings IEEE conference on decision and control, San Antonio, TX (pp. 1326¯1329).

Sontag, E. D. (1995). From linear to nonlinear: Some complexity comparisons. Proceedings of the IEEE conference decision and control, New Orleans (pp. 2916¯2920).

Sontag, E. D. (1996a). Interconnected automata and linear systems: A theoretical framework in discrete-time. In R. Alur, T. Henzinger, & E. D. Sontag, Hybrid systems III: Verification and control (pp. 436¯448). Berlin: Springer.

Sontag, E. D. (1996b). Recurrent neural networks: Some learning and systems-theoretic aspects. Preprint.

Syrmos, V., Abdallah, C. and Dorato, P., 1994. Static output feedback: A survey. Automatica 33, pp. 125¯137.

Tarski, A., 1951. A decision method for elementary algebra and geometry, University of California Press, Berkeley, USA.

Tempo, R., Bai, E.W., & Dabbene, F. (1996). Probabilistic robustness analysis: Explicit bounds for the minimum number of sampling points. Proceedings on IEEE conference decision and control (pp. 2412¯2420).

Toker, O., 1997. On the complexity of the robust stability problem for linear parameter varying systems. Automatica 33, pp. 2015¯2017. Abstract | Journal Format-PDF (268 K) | MathSciNet

Toker, O. (1996). On the algorithmic unsolvability of some stability problems for discrete event systems. Proceedings of IFAC world congress (pp. 353¯358).

Toker, O., & de Jager, P. (1999). Conservatism of the standard upper bound test. In V. D. Blondel, E. D. Sontag, M. Vidyasagar, & J. C. Willems, Open problems in mathematical systems and control theory, Problem 20. London: Springer.

Toker, O. and Ozbay, H., 1998. On the complexity of purely complex computation and related problems in multidimensional systems. IEEE Transactions on Automatic Control 43, pp. 409¯414. Abstract  | MathSciNet

Toker, O. and Ozbay, H., 1996. Complexity issues in robust stability of linear delay-differential systems. Mathematics of Control, Signals, and Systems 9, pp. 386¯400. Abstract  | MathSciNet

Toker, O., & Ozbay, H. (1995). On the NP-hardness of solving bilinear matrix inequalities and simultaneous stabilization with static output feedback. Proceedings of the IEEE American control conference, Seattle, WA (pp. 2525¯2526).

Traub, J., Wasilkowski, G. and Woniakowski, H., 1988. Information-based complexity, Academic Press, New York.

Trei, S. (1999). The gap between complex structured singular value and its upper bound is infinite. Preprint, http://www.math.msu.edu/~treil/research.html.

Tseng, P., 1990. Solving h-horizon, stationary Markov decision problems in time proportional to log h. Operations Research Letters 9, pp. 287¯297. INSPEC Compendex  | MathSciNet

Tsitsiklis, J. N. (1984). Problems in decentralized decision making and computation. Ph.D. thesis, Department of Electrical Engineering and Computer Science. Cambridge, MA: MIT Press.

Tsitsiklis, J.N., 1989. On the control of discrete-event dynamical systems. Mathematics of Control, Signals, and Systems 2, pp. 95¯107. INSPEC Compendex  | MathSciNet

Tsitsiklis, J.N. and Athans, M., 1985. On the complexity of decentralized decision making and detection problems. IEEE Transactions on Automatic Control 30, pp. 440¯446. INSPEC Compendex  | MathSciNet

Tsitsiklis, J.N. and Blondel, V.D., 1997. The Lyapunov exponent and joint spectral radius of pairs of matrices are hard ¯¯ when not impossible ¯¯ to compute and to approximate. Mathematics of Control Signals, and Systems 10, pp. 31¯40. Abstract  | MathSciNet

Tsitsiklis, J.N. and Blondel, V.D., 1997. Lyapunov exponents of pairs of matrices, a correction. Mathematics of Control, Signals, and Systems 10, p. 381.

Vidyasagar, M., 1998. A theory of learning and generalization. With applications to neural networks and control systems, Springer, London.

Vidyasagar, M., & Blondel, V. D. (1999). Probabilistic solutions to some NP-hard matrix problems. Preprint.

Whitt, W., 1978. Approximations of dynamic programs ¯¯ I. Mathematics of Operations Research 3, pp. 231¯243. INSPEC Compendex  | MathSciNet

Whitt, W., 1978. Approximations of dynamic programs ¯¯ II. Mathematics of Operations Research 4, pp. 179¯185.

Whittle, P. (1988). Restless bandits: Activity allocation in a changing world. In J. Gani, A celebration of applied probability, Journal of Applied Probability, 25A, 287¯298.

Witsenhausen, H.S., 1968. A counterexample in stochastic optimum control. SIAM Journal of Control and Optimization 6, pp. 138¯147.

Zhou, K., Doyle, J.C. and Glover, K., 1995. Robust and optimal control, Prentice Hall, Englewood Cliffs, NJ.

Zwick, U. and Paterson, M., 1996. The complexity of mean payoff games on graphs. Theoretical Computer Science 158, pp. 343¯359. Abstract | Journal Format-PDF (835 K) | MathSciNet

*1 This paper was not presented at any IFAC meeting. This paper was recommended for publication in revised from by Editor M. Morari. This work was supported by the NATO under grant CRG-961115, by the European Community Framework IV program through the research network ALAPEDES, by the AFOSR under grant F49620-99-10320 and by the ARO under the MURI grant DAAH04-96-1-0341. It was partially carried out while the second author was visiting the EECS department at the University of California at Berkeley.

Corresponding author. Tel.: +32-10-472381; fax: +32-10-472180; email: blondel@inma.ucl.ac.be

1 We use the O(·) notation, whose precise meaning is the following. Let f and g be functions that map positive numbers to positive numbers. We write f (n)=O(g(n)) if there exist positive numbers n0 and c such that f (n)cg(n) for all nn0.

2 A word from a given alphabet (e.g., the alphabet {a,b}) is a concatenation of finitely many symbols from that alphabet (e.g., aababb), or the "empty" word. For any two words s and , s stands for their concatenation.

3 This definition of stability relates to continuous-time systems. We restrict the discussion in 3.1, 3.2, 3.3 and 3.4 to this continuous-time setting.

4 A literal is either a Boolean variable ui or its negation . A one-literal clause consists of a single literal. A two-literal clause is a disjunction (xy) of two literals x and y.

5 The results presented in Poljak and Rohn (1993) and Nemirovskii (1993) were found independently. Poljak and Rohn (1993) was submitted about two years before Nemirovskii (1993), but for editorial reasons, the two papers appeared in the same issue of Mathematics of Control, Signals, and Systems.

6 In the problem, we are given integers a1,...,an, whose sum is some number K and we are asked whether there exists a subset of these integers whose sum is K/2.

7 This formulation has been suggested by M. Overton (personal communication).

8 This conjecture can be restated as saying that the convergence to zero of all periodic products of a given finite set of matrices, implies the same for all possible products.

9 In fact, a careful examination of the proof in that reference shows that a stronger result is true, ruling out algorithms that run in time which is polynomial in 1/.

10 This result is stated incorrectly in Tsitsiklis and Blondel (1997a); a correction appears in Tsitsiklis and Blondel (1997b).

11 A universal Turing machine is a Turing machine M that receives as input the description of another machine M' and a string s, and then simulates the operation of machine M' on input s.

12 Notice that the title of the reference Hyotyniemu (1997) involves the term "stability", but it is actually used in a sense different than the usual one in systems theory.

13 In high-order neural nets, activations are combined using multiplication as opposed to just linear combinations.

14 The example in Melekopoglou and Condon (1994) involves an undiscounted problem, but as pointed out in Littman et al. (1995), it applies to discounted problems as well.

15 Although these results are stated for the problem of uniform approximation, the proof of the lower bounds also applies to the problem of approximating J* at a particular given state.

16 This counterexample showed that optimal controllers for LQG problems with a nonclassical information structure are not necessarily linear, thus eliminating the hope of extending LQG theory to decentralized control.

Vitae

Blondel: John N. Tsitsiklis was born in Thessaloniki, Greece, in 1958. He received the B.S. degree in Mathematics (1980), and the B.S. (1980), M.S. (1981) and Ph.D. (1984) degrees in Electrical Engineering, all from the Massachusetts Institute of Technology, Cambridge, Massachusetts.

During the academic year 1983¯1984, he was an acting assistant professor of Electrical Engineering at Stanford University, Stanford, California. Since 1984, he has been with the Massachusetts Institute of Technology, where he is currently Professor of Electrical Engineering and Computer Science. He has served as acting co-director of the MIT Laboratory for Information and Decision Systems (Spring 1996 and 1997). He has also been a visitor with the Dept. of EECS at the University of California at Berkeley, and the Institute for Computer Science in Iraklion, Greece. His research interests are in the fields of systems, optimization, control, and operations research. He has written about 80 journal papers in these areas.

He is a coauthor of "Parallel and Distributed Computation: Numerical Methods" (with D. Bertsekas, 1989), "Neuro-Dynamic Programming" (with Dimitri Bertsekas, 1996), and "Introduction to Linear Optimization (with Dimitris Bertsimas, 1997). He has been a recipient of an IBM Faculty Development Award (1983), an NSF Presidential Young Investigator Award (1986), an Outstanding Paper Award by the IEEE Control Systems Society, the M.I.T. Edgerton Faculty Achievement Award (1989), the Bodossakis Foundation Prize (1995), the INFORMS/CSTS prize (1997), and is a Fellow of the IEEE (1999). He is an associate editor of Applied Mathematics Letters and has been an associate editor of the IEEE Transactions on Automatic Control and Automatica.

For a photograph of J.N. Tsitsiklis, see Automatica 35(3), March (1999).

Tsitsiklis: Vincent D. Blondel received a first degree in Applied Mathematics from the University of Louvain (Louvain-la-Neuve, Belgium) in 1988, a Master Degree in Pure Mathematics from Imperial College (London) and a Ph.D. in 1992 from the University of Louvain.

In 1993 he was visiting scientist at Oxford University. During the academic year 1993¯1994 he was Goran Gustafsson Fellow at the Department of Mathematics of the Royal Institute of Technology (Stockholm). In 1993¯1994 he was a research fellow at the French National Research Center in Control and Computer Science INRIA (Paris). From 1994 to 1999 he was Associate Professor of Mathematics at the University of Liège (Belgium). He has recently moved to the University of Louvain where he is currently Professor of Applied Mathematics. He has been a visitor with the Australian National University, the Massachusetts Institute of Technology, the University of Paris VII and the Ecole Normale Supérieure in Lyon.

Dr. Blondel's major current research interests lie in several areas of mathematical systems theory and theoretical computer science. He is the author of "Simultaneous Stabilization of Linear Systems" (1994) and is one of the editors of "Open Problems in Mathematical Systems Theory" (1999). He has been a recipient of a grant from the Trustees of the Mathematics Institute of Oxford University (1992), the Prize Agathon De Potter of the Belgian Royal Academy of Science (1993) and the Prize Paul Dubois of the Montefiore Institute (1993). He is the coordinator of a NATO collaborative grant with the Massachusetts Institute of Technology and is a partner in a European TRM network on discrete event systems. He is an associate editor of Systems and Control Letters, of the European Journal of Control and of Mathematics of Control, Signals, and Systems.

For a photograph of V.D. Blondel, see Automatica 35(3), March (1999).


Automatica
SummaryPlus
Article
Journal Format-PDF (293 K)
Volume 36, Issue 9
September 2000
Pages 1249-1274


2 of 9Article ListPreviousNext
HomeJournalsSearchAlertsAccountHelp

Send feedback to ScienceDirect
Software and compilation © 2000 ScienceDirect. All rights reserved.
ScienceDirect® is an Elsevier Science B.V. registered trademark.