book_name
stringclasses
983 values
chunk_index
int64
0
503
text
stringlengths
1.01k
10k
1012_(GTM178)Nonsmooth Analysis and Control Theory
0
# GraduateTexts inMathematics F.H. Clarke Yu.S. Ledyaev R.J. Stern P.R. Wolenski # Nonsmooth Analysis and Control Theory Springer ## Graduate Texts in Mathematics 178 Editorial Board S. Axler F.W. Gehring K.A. Ribet Springer New York Berlin Heidelberg Barcelona Budapest Hong Kong London Milan Paris Santa Clara Singapore Tokyo ## Graduate Texts in Mathematics 1 TAKEUTI/ZARING. Introduction to Axiomatic Set Theory. 2nd ed. 2 Oxtoby. Measure and Category. 2nd ed. 3 Schaefer. Topological Vector Spaces. 4 Hilton/Stammbach. A Course in Homological Algebra. 2nd ed. 5 MAC LANE. Categories for the Working Mathematician. 6 Hughes/Piper. Projective Planes. 7 Serre. A Course in Arithmetic. 8 TAKEUTI/ZARING. Axiomatic Set Theory. 9 Humphreys. Introduction to Lie Algebras and Representation Theory. 10 COHEN. A Course in Simple Homotopy Theory. 11 Conway. Functions of One Complex Variable I. 2nd ed. 12 Beals. Advanced Mathematical Analysis. 13 Anderson/Fuller. Rings and Categories of Modules. 2nd ed. 14 Golubitsky/Guillemin. Stable Mappings and Their Singularities. 15 Berberian. Lectures in Functional Analysis and Operator Theory. 16 WINTER. The Structure of Fields. 17 ROSENBLATT. Random Processes. 2nd ed. 18 Halmos. Measure Theory. 19 Halmos. A Hilbert Space Problem Book. 2nd ed. 20 Husemoller. Fibre Bundles. 3rd ed. 21 Humphreys. Linear Algebraic Groups. 22 Barnes/Mack. An Algebraic Introduction to Mathematical Logic. 23 Greub. Linear Algebra. 4th ed. 24 Holmes. Geometric Functional Analysis and Its Applications. 25 Hewitt/Stromberg. Real and Abstract Analysis. 26 Manes. Algebraic Theories. 27 Kelley. General Topology. 28 Zariski/Samuel. Commutative Algebra. Vol.I. 29 Zariski/Samuel. Commutative Algebra. Vol.II. 30 JACOBSON. Lectures in Abstract Algebra I. Basic Concepts. 31 JACOBSON. Lectures in Abstract Algebra II. Linear Algebra. 32 JACOBSON. Lectures in Abstract Algebra III. Theory of Fields and Galois Theory. 33 Hirsch. Differential Topology. 34 Spitzer. Principles of Random Walk. 2nd ed. 35 ALEXANDER/WERMER. Several Complex Variables and Banach Algebras. 3rd ed. 36 Kelley/Namioka et al. Linear Topological Spaces. 37 Monk. Mathematical Logic. 38 Grauert/Fritzsche. Several Complex Variables. 39 Arveson. An Invitation to \( {C}^{ * } \) -Algebras. 40 Kemeny/Snell/Knapp. Denumerable Markov Chains. 2nd ed. 41 APOSTOL. Modular Functions and Dirichlet Series in Number Theory. 2nd ed. 42 SERRE. Linear Representations of Finite Groups. 43 GILLMAN/JERISON. Rings of Continuous Functions. 44 KENDIG. Elementary Algebraic Geometry. 45 Loève. Probability Theory I. 4th ed. 46 Loève. Probability Theory II. 4th ed. 47 Morse. Geometric Topology in Dimensions 2 and 3. 48 SACHS/WU. General Relativity for Mathematicians. 49 Gruenberg/Weir. Linear Geometry. 2nd ed. 50 Edwards. Fermat's Last Theorem. 51 KLINGENBERG. A Course in Differential Geometry. 52 Hartshorne. Algebraic Geometry. 53 Manin. A Course in Mathematical Logic. 54 Graver/Watkins. Combinatorics with Emphasis on the Theory of Graphs. 55 Brown/Pearcy. Introduction to Operator Theory I: Elements of Functional Analysis. 56 Massey. Algebraic Topology: An Introduction. 57 Crowell/Fox. Introduction to Knot Theory. 58 KOBLITZ. \( p \) -adic Numbers, \( p \) -adic Analysis, and Zeta-Functions. 2nd ed. 59 LANG. Cyclotomic Fields. 60 Arnold. Mathematical Methods in Classical Mechanics. 2nd ed. F.H. Clarke Yu.S. Ledyaev R.J. Stern P.R. Wolenski # Nonsmooth Analysis and Control Theory F.H. Clarke Yu.S. Ledyaev Institut Desargues Steklov Mathematics Institute Université de Lyon I Moscow, 117966 Villeurbanne, 69622 Russia France R.J. Stern P.R. Wolenski Department of Mathematics Department of Mathematics Concordia University Louisiana State University 7141 Sherbrooke St. West Baton Rouge, LA 70803-0001 Montreal, PQ H4B 1R6 USA Canada Editorial Board S. Axler F. W. Gehring K.A. Ribet Mathematics Department Mathematics Department Department of Mathematics San Francisco State East Hall University of California University University of Michigan at Berkeley San Francisco, CA 94132 Ann Arbor, MI 48109 Berkeley, CA 94720-3840 USA USA USA ## Mathematics Subject Classification (1991): 49J52, 58C20, 90C48 With 8 figures. Library of Congress Cataloging-in-Publication Data Nonsmooth analysis and control theory / F.H. Clarke . . . [et al.]. p. cm. - (Graduate texts in mathematics ; 178) Includes bibliographical references and index. ISBN 0-387-98336-8 (hardcover : alk. paper) 1. Control Theory. 2. Nonsmooth optimization. I. Clarke, Francis H. II. Series. QA402.3.N66 1998 \( {515}^{\prime }{.64} - \mathrm{{dc}}{21} \) 97-34140 (C) 1998 Springer-Verlag New York, Inc. All rights reserved. This work may not be translated or copied in whole or in part without the written permission of the publisher (Springer-Verlag New York, Inc., 175 Fifth Avenue, New York, NY 10010, USA), except for brief excerpts in connection with reviews or scholarly analysis. Use in connection with any form of information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed is forbidden. The use of general descriptive names, trade names, trademarks, etc., in this publication, even if the former are not especially identified, is not to be taken as a sign that such names, as understood by the Trade Marks and Merchandise Marks Act, may accordingly be used freely by anyone. The authors dedicate this book: to Gail, Julia, and Danielle; to Sofia, Simeon, and Irina; to Judy, Adam, and Sach; and to Mary and Anna. ## Preface Pardon me for writing such a long letter; I had not the time to write a short one. —Lord Chesterfield Nonsmooth analysis refers to differential analysis in the absence of differentiability. It can be regarded as a subfield of that vast subject known as nonlinear analysis. While nonsmooth analysis has classical roots (we claim to have traced its lineage back to Dini), it is only in the last decades that the subject has grown rapidly. To the point, in fact, that further development has sometimes appeared in danger of being stymied, due to the plethora of definitions and unclearly related theories. One reason for the growth of the subject has been, without a doubt, the recognition that nondifferentiable phenomena are more widespread, and play a more important role, than had been thought. Philosophically at least, this is in keeping with the coming to the fore of several other types of irregular and nonlinear behavior: catastrophes, fractals, and chaos. In recent years, nonsmooth analysis has come to play a role in functional analysis, optimization, optimal design, mechanics and plasticity, differential equations (as in the theory of viscosity solutions), control theory, and, increasingly, in analysis generally (critical point theory, inequalities, fixed point theory, variational methods ...). In the long run, we expect its methods and basic constructs to be viewed as a natural part of differential analysis. We have found that it would be relatively easy to write a very long book on nonsmooth analysis and its applications; several times, we did. We have now managed not to do so, and in fact our principal claim for this work is that it presents the essentials of the subject clearly and succinctly, together with some of its applications and a generous supply of interesting exercises. We have also incorporated in the text a number of new results which clarify the relationships between the different schools of thought in the subject. We hope that this will help make nonsmooth analysis accessible to a wider audience. In this spirit, the book is written so as to be used by anyone who has taken a course in functional analysis. We now proceed to discuss the contents. Chapter 0 is an Introduction in which we allow ourselves a certain amount of hand-waving. The intent is to give the reader an avant-goût of what is to come, and to indicate at an early stage why the subject is of interest. There are many exercises in Chapters 1 to 4, and we recommend (to the active reader) that they be done. Our experience in teaching this material has had a great influence on the writing of this book, and indicates that comprehension is proportional to the exercises done. The end-of-chapter problems also offer scope for deeper understanding. We feel no guilt in calling upon the results of exercises later as needed. Chapter 1, on proximal analysis, should be done carefully by every reader of this book. We have chosen to work here in a Hilbert space, although the greater generality of certain Banach spaces having smooth norms would be another suitable context. We believe the Hilbert space setting makes for a more accessible theory on first exposure, while being quite adequate for later applications. Chapter 2 is devoted to the theory of generalized gradients, which constitutes the other main approach (other than proximal) to developing non-smooth analysis. The natural habitat of this theory is Banach space, which is the choice made. The relationship between these two principal approaches is now well understood, and is clearly delineated here. As for the preceding chapter, the treatment is not encyclopedic, but covers the important ideas. In Chapter 3 we develop certain special topics, the first of which is value function analysis for constrained optimization. This topic is previewed in Chapter 0, and §3.1 is helpful, though not essential, in understanding certain proofs in the latter part of Chapter 4. The next topic, mean value inequalities, offers a glimpse of more advanced calculus. It also serves as a basis for the solvability results of the next section, which features the Graves-Lyusternik Theorem and the Lipschitz Inverse Function Theorem. Section 3.4 is a brief look at a third route to nonsmooth calculus, one that bases itself upon dir
1012_(GTM178)Nonsmooth Analysis and Control Theory
1
h constitutes the other main approach (other than proximal) to developing non-smooth analysis. The natural habitat of this theory is Banach space, which is the choice made. The relationship between these two principal approaches is now well understood, and is clearly delineated here. As for the preceding chapter, the treatment is not encyclopedic, but covers the important ideas. In Chapter 3 we develop certain special topics, the first of which is value function analysis for constrained optimization. This topic is previewed in Chapter 0, and §3.1 is helpful, though not essential, in understanding certain proofs in the latter part of Chapter 4. The next topic, mean value inequalities, offers a glimpse of more advanced calculus. It also serves as a basis for the solvability results of the next section, which features the Graves-Lyusternik Theorem and the Lipschitz Inverse Function Theorem. Section 3.4 is a brief look at a third route to nonsmooth calculus, one that bases itself upon directional subderivates. It is shown that the salient points of this theory can be derived from the earlier results. We also present here a self-contained proof of Rademacher’s Theorem. In \( §{3.5} \) we develop some machinery that is used in the following chapter, notably measurable selection. We take a quick look at variational functionals, but by-and-large, the calculus of variations has been omitted. The final section of the chapter examines in more detail some questions related to tangency. Chapter 4, as its title implies, is a self-contained introduction to the theory of control of ordinary differential equations. This is a biased introduction, since one of its avowed goals is to demonstrate virtually all of the preceding theory in action. It makes no attempt to address issues of modeling or of implementation. Nonetheless, most of the central issues in control are studied, and we believe that any serious student of mathematical control theory will find it essential to have a grasp of the tools that are developed here via nonsmooth analysis: invariance, viability, trajectory monotonicity, viscosity solutions, discontinuous feedback, and Hamiltonian inclusions. We believe that the unified and geometrically motivated approach presented here for the first time has merits that will continue to make themselves felt in the subject. We now make some suggestions for the reader who does not have the time to cover all of the material in this book. If control theory is of less interest, then Chapters 1 and 2, together with as much of Chapter 3 as time allows, constitutes a good introduction to nonsmooth analysis. At the other extreme is the reader who wishes to do Chapter 4 virtually in its entirety. In that case, a jump to Chapter 4 directly after Chapter 1 is feasible; only occasional references to material in Chapters 2 and 3 is made, up to \( §{4.8} \) , and in such a way that the reader can refer back without difficulty. The two final sections of Chapter 4 have a greater dependence on Chapter 2, but can still be covered if the reader will admit the proofs of the theorems. A word on numbering. All items are numbered in sequence within a section; thus Exercise 7.2 precedes Theorem 7.3, which is followed by Corollary 7.4. For references between two chapters, an extra initial digit refers to the chapter number. Thus a result that would be referred to as Theorem 7.3 within Chapter 1 would be invoked as Theorem 1.7.3 from within Chapter 4. All equation numbers are simple, as in (3), and start again at (1) at the beginning of each section (thus their effect is only local). A reference to \( §3 \) is to the third section of the current chapter, while \( §{2.3} \) refers to the third section of Chapter 2. A summary of our notational conventions is given in \( §{0.5} \), and a Symbol Glossary appears in the Notes and Comments at the end of the book. We would like to express our gratitude to the personnel of the Centre de Recherches Mathématiques (CRM) of l’Université de Montréal, and in particular to Louise Letendre, for their invaluable help in producing this book. Finally, we learned as the book was going to press, of the death of our friend and colleague Andrei Subbotin. We wish to express our sadness at his passing, and our appreciation of his many contributions to our subject. Francis Clarke, Lyon Yuri Ledyaev, Moscow Ron Stern, Montréal Peter Wolenski, Baton Rouge May 1997 ## Contents Preface vii List of Figures xiii 0 Introduction 1 1 Analysis Without Linearization 1 2 Flow-Invariant Sets 7 3 Optimization 10 4 Control Theory 15 5 Notation 18 1 Proximal Calculus in Hilbert Space 21 1 Closest Points and Proximal Normals . 21 2 Proximal Subgradients 27 3 The Density Theorem 39 4 Minimization Principles 43 5 Quadratic Inf-Convolutions 44 6 The Distance Function 47 7 Lipschitz Functions 51 8 The Sum Rule 54 9 The Chain Rule 58 10 Limiting Calculus 61 11 Problems on Chapter 1 63 xii Contents 2 Generalized Gradients in Banach Space 69 1 Definition and Basic Properties 69 2 Basic Calculus 74 3 Relation to Derivatives 78 4 Convex and Regular Functions 80 5 Tangents and Normals 83 6 Relationship to Proximal Analysis 88 7 The Bouligand Tangent Cone and Regular Sets 90 8 The Gradient Formula in Finite Dimensions 93 9 Problems on Chapter 2 . 96 3 Special Topics 103 1 Constrained Optimization and Value Functions 103 2 The Mean Value Inequality 111 3 Solving Equations 125 4 Derivate Calculus and Rademacher's Theorem 136 5 Sets in \( {L}^{2} \) and Integral Functionals 148 6 Tangents and Interiors 165 7 Problems on Chapter 3 . 170 4 A Short Course in Control Theory 177 1 Trajectories of Differential Inclusions 177 2 Weak Invariance 188 3 Lipschitz Dependence and Strong Invariance 195 4 Equilibria 202 5 Lyapounov Theory and Stabilization 208 6 Monotonicity and Attainability 215 7 The Hamilton-Jacobi Equation and Viscosity Solutions 222 8 Feedback Synthesis from Semisolutions 228 9 Necessary Conditions for Optimal Control 230 10 Normality and Controllability 244 11 Problems on Chapter 4 . 247 Notes and Comments 257 List of Notation 263 Bibliography 265 Index 273 List of Figures 0.1 Torricelli's table. 12 0.2 Discontinuity of the local projection. 13 1.1 A set \( S \) and some of its boundary points. 22 1.2 A point \( {x}_{1} \) and its five projections. 24 1.3 The epigraph of a function. 30 \( {1.4}\;\zeta \) belongs to \( {\partial }_{P}f\left( x\right) \) . 35 4.1 The set \( S \) of Exercise 2.12. 195 4.2 The set \( S \) of Exercise 4.3. 204 0 Introduction Experts are not supposed to read this book at all. -R.P. Boas, A Primer of Real Functions We begin with a motivational essay that previews a few issues and several techniques that will arise later in this book. ## 1 Analysis Without Linearization Among the issues that routinely arise in mathematical analysis are the following three: - to minimize a function \( f\left( x\right) \) ; - to solve an equation \( F\left( x\right) = y \) for \( x \) as a function of \( y \) ; and - to derive the stability of an equilibrium point \( {x}^{ * } \) of a differential equation \( \dot{x} = \varphi \left( x\right) \) . None of these issues imposes by its nature that the function involved \( (f \) , \( F \), or \( \varphi \) ) be smooth (differentiable); for example, we can reasonably aim to minimize a function which is merely continuous, if growth or compactness is postulated. Nonetheless, the role of derivatives in questions such as these has been central, due to the classical technique of linearization. This term refers to the construction of a linear local approximation of a function by means of its derivative at a point. Of course, this approach requires that the derivative exists. When applied to the three scenarios listed above, linearization gives rise to familiar and useful criteria: - at a minimum \( x \), we have \( {f}^{\prime }\left( x\right) = 0 \) (Fermat’s Rule); - if the \( n \times n \) Jacobian matrix \( {F}^{\prime }\left( x\right) \) is nonsingular, then \( F\left( x\right) = y \) is locally invertible (the Inverse Function Theorem); and - if the eigenvalues of \( {\varphi }^{\prime }\left( {x}^{ * }\right) \) have negative real parts, the equilibrium is locally stable. The main purpose of this book is to introduce and motivate a set of tools and methods that can be used to address these types of issues, as well as others in analysis, optimization, and control, when the underlying data are not (necessarily) smooth. In order to illustrate in a simple setting how this might be accomplished, and in order to make contact with what could be viewed as the first theorem in what has become known as nonsmooth analysis, let us consider the following question: to characterize in differential, thus local terms, the global property that a given continuous function \( f : \mathbb{R} \rightarrow \mathbb{R} \) is decreasing (i.e., \( x \leq y \Rightarrow f\left( y\right) \leq f\left( x\right) \) ). If the function \( f \) admits a continuous derivative \( {f}^{\prime } \), then the integration formula \[ f\left( y\right) = f\left( x\right) + {\int }_{x}^{y}{f}^{\prime }\left( t\right) {dt} \] leads to a sufficient condition for \( f \) to be decreasing: that \( {f}^{\prime }\left( t\right) \) be nonpositive for each \( t \) . It is easy to see that this is necessary as well, so a satisfying characterization via \( {f}^{\prime } \) is obtained. If we go beyond the class of continuously differentiable functions, the situation becomes much more complex. It is known, for example, that there exists a strictly decreasing continuous \( f \) for which we have \( {f}^{\prime }\left( t\right) = 0 \) almost everywhere. For such a function, the derivative appears to fail us, insofar as characterizing decrease is concerned. In 1878, Ulysse Dini introduced certain constructs, one of which is the following (lower, right) d
1012_(GTM178)Nonsmooth Analysis and Control Theory
2
( x \leq y \Rightarrow f\left( y\right) \leq f\left( x\right) \) ). If the function \( f \) admits a continuous derivative \( {f}^{\prime } \), then the integration formula \[ f\left( y\right) = f\left( x\right) + {\int }_{x}^{y}{f}^{\prime }\left( t\right) {dt} \] leads to a sufficient condition for \( f \) to be decreasing: that \( {f}^{\prime }\left( t\right) \) be nonpositive for each \( t \) . It is easy to see that this is necessary as well, so a satisfying characterization via \( {f}^{\prime } \) is obtained. If we go beyond the class of continuously differentiable functions, the situation becomes much more complex. It is known, for example, that there exists a strictly decreasing continuous \( f \) for which we have \( {f}^{\prime }\left( t\right) = 0 \) almost everywhere. For such a function, the derivative appears to fail us, insofar as characterizing decrease is concerned. In 1878, Ulysse Dini introduced certain constructs, one of which is the following (lower, right) derivate: \[ {Df}\left( x\right) \mathrel{\text{:=}} \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{f\left( {x + t}\right) - f\left( x\right) }{t}. \] Note that \( {Df}\left( x\right) \) can equal \( + \infty \) or \( - \infty \) . It turns out that \( {Df} \) will serve our purpose, as we now see. 1.1. Theorem. The continuous function \( f : \mathbb{R} \rightarrow \mathbb{R} \) is decreasing iff \[ {Df}\left( x\right) \leq 0\forall x \in \mathbb{R}. \] Although this result is well known, and in any case greatly generalized in a later chapter, let us indicate a nonstandard proof of it now, in order to bring out two themes that are central to this book: optimization and nonsmooth calculus. Note first that \( {Df}\left( x\right) \leq 0 \) is an evident necessary condition for \( f \) to be decreasing, so it is the sufficiency of this property that we must prove. Let \( x, y \) be any two numbers with \( x < y \) . We will prove that for any \( \delta > 0 \) , we have \[ \min \{ f\left( t\right) : y \leq t \leq y + \delta \} \leq f\left( x\right) . \] (1) This implies \( f\left( y\right) \leq f\left( x\right) \), as required. As a first step in the proof of (1), let \( g \) be a function defined on \( \left( {x - \delta, y + \delta }\right) \) with the following properties: (a) \( g \) is continuously differentiable, \( g\left( t\right) \geq 0, g\left( t\right) = 0 \) iff \( t = y \) ; (b) \( {g}^{\prime }\left( t\right) < 0 \) for \( t \in \left( {x - \delta, y}\right) \) and \( {g}^{\prime }\left( t\right) \geq 0 \) for \( t \in \lbrack y, y + \delta ) \) ; and (c) \( g\left( t\right) \rightarrow \infty \) as \( t \downarrow x - \delta \), and also as \( t \uparrow y + \delta \) . It is easy enough to give an explicit formula for such a function; we will not do so. Now consider the minimization over \( \left( {x - \delta, y + \delta }\right) \) of the function \( f + g \) ; by continuity and growth, the minimum is attained at a point \( z \) . A necessary condition for a local minimum of a function is that its Dini derivate be nonnegative there, as is easily seen. This gives \[ D\left( {f + g}\right) \left( z\right) \geq 0. \] Because \( g \) is smooth, we have the following fact (in nonsmooth calculus!): \[ D\left( {f + g}\right) \left( z\right) = {Df}\left( z\right) + {g}^{\prime }\left( z\right) . \] Since \( {Df}\left( z\right) \leq 0 \) by assumption, we derive \( {g}^{\prime }\left( z\right) \geq 0 \), which implies that \( z \) lies in the interval \( \lbrack y, y + \delta ) \) . We can now estimate the left side of (1) as follows: \[ \min \{ f\left( t\right) : y \leq t \leq y + \delta \} \leq f\left( z\right) \] \[ \leq f\left( z\right) + g\left( z\right) \text{ (since }g \geq 0\text{ ) } \] \[ \leq f\left( x\right) + g\left( x\right) \text{ (since }z\text{ minimizes }f + g\text{ ). } \] We now observe that the entire argument to this point will hold if \( g \) is replaced by \( {\varepsilon g} \), for any positive number \( \varepsilon \) (since \( {\varepsilon g} \) continues to satisfy the listed properties for \( g \) ). This observation implies (1) and completes the proof. We remark that the proof of Theorem 1.1 will work just as well if \( f \), instead of being continuous, is assumed to be lower semicontinuous, which is the underlying hypothesis made on the functions that appear in Chapter 1. An evident corollary of Theorem 1.1 is that a continuous everywhere differentiable function \( f \) is decreasing iff its derivative \( {f}^{\prime }\left( x\right) \) is always nonpositive, since when \( {f}^{\prime }\left( x\right) \) exists it coincides with \( {Df}\left( x\right) \) . This could also be proved directly from the Mean Value Theorem, which asserts that when \( f \) is differentiable we have \[ f\left( y\right) - f\left( x\right) = {f}^{\prime }\left( z\right) \left( {y - x}\right) \] for some \( z \) between \( x \) and \( y \) . ## Proximal Subgradients We will now consider monotonicity for functions of several variables. When \( x, y \) are points in \( {\mathbb{R}}^{n} \), the inequality \( x \leq y \) will be understood in the component-wise sense: \( {x}_{i} \leq {y}_{i} \) for \( i = 1,2,\ldots, n \) . We say that a given function \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is decreasing provided that \( f\left( y\right) \leq f\left( x\right) \) whenever \( x \leq y \) . Experience indicates that the best way to extend Dini's derivates to functions of several variables is as follows: for a given direction \( v \) in \( {\mathbb{R}}^{n} \) we define \[ {Df}\left( {x;v}\right) \mathrel{\text{:=}} \mathop{\liminf }\limits_{\substack{{t \downarrow 0} \\ {w \rightarrow v} }}\frac{f\left( {x + {tw}}\right) - f\left( x\right) }{t}. \] We call \( {Df}\left( {x;v}\right) \) a directional subderivate. Let \( {\mathbb{R}}_{ + }^{n} \) denote the positive orthant in \( {\mathbb{R}}^{n} \) : \[ {\mathbb{R}}_{ + }^{n} \mathrel{\text{:=}} \left\{ {x \in {\mathbb{R}}^{n} : x \geq 0}\right\} \] We omit the proof of the following extension of Theorem 1.1, which can be given along the lines of that of Theorem 1.1. 1.2. Theorem. The continuous function \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is decreasing iff \( {Df}\left( {x;v}\right) \leq 0\forall x \) in \( {\mathbb{R}}^{n},\forall v \in {\mathbb{R}}_{ + }^{n} \) . When \( f \) is continuously differentiable, it is the case that \( {Df}\left( {x;v}\right) \) agrees with \( \langle \nabla f\left( x\right), v\rangle \), an observation that leads to the following consequence of the theorem: 1.3. Corollary. A continuously differentiable function \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is decreasing iff \( \nabla f\left( x\right) \leq 0\forall x \in {\mathbb{R}}^{n} \) . Since it is easier in principle to examine one gradient vector than an infinite number of directional subderivates, we are led to seek an object that could replace \( \nabla f\left( x\right) \) in a result such as Corollary 1.3, when \( f \) is nondifferentiable. A concept that turns out to be a powerful tool in characterizing a variety of functional properties is that of the proximal subgradient. A vector \( \zeta \) in \( {\mathbb{R}}^{n} \) is said to be a proximal subgradient of \( f \) at \( x \) provided that there exist a neighborhood \( U \) of \( x \) and a number \( \sigma > 0 \) such that \[ f\left( y\right) \geq f\left( x\right) + \langle \zeta, y - x\rangle - \sigma \parallel y - x{\parallel }^{2}\forall y \in U. \] The set of such \( \zeta \), if any, is denoted \( {\partial }_{P}f\left( x\right) \) and is referred to as the proximal subdifferential. The existence of a proximal subgradient \( \zeta \) at \( x \) corresponds to the possibility of approximating \( f \) from below (thus in a one-sided manner) by a function whose graph is a parabola. The point \( \left( {x, f\left( x\right) }\right) \) is a contact point between the graph of \( f \) and the parabola, and \( \zeta \) is the slope of the parabola at that point. Compare this with the usual derivative, in which the graph of \( f \) is approximated by an affine function. Among the many properties of \( {\partial }_{P}f \) developed later will be a Mean Value Theorem asserting that for given points \( x \) and \( y \), for any \( \varepsilon > 0 \), we have \[ f\left( y\right) - f\left( x\right) \leq \langle \zeta, y - x\rangle + \varepsilon \] where \( \zeta \) belongs to \( {\partial }_{P}f\left( z\right) \) for some point \( z \) which lies within \( \varepsilon \) of the line segment joining \( x \) and \( y \) . This theorem requires of \( f \) merely lower semicontinuity. A consequence of this is the following. 1.4. Theorem. A lower semicontinuous function \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is decreasing iff \( \zeta \leq 0\forall \zeta \) in \( {\partial }_{P}f\left( x\right) ,\forall x \) in \( {\mathbb{R}}^{n} \) . We remark that Theorem 1.4 subsumes Theorem 1.2, as a consequence of the following implication, which the reader may readily confirm: \[ \zeta \in {\partial }_{P}f\left( x\right) \Rightarrow {Df}\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v. \] While characterizations such as the one given by Theorem 1.4 are of intrinsic interest, it is reassuring to know that they can be and have been of actual use in practice. For example, in developing an existence theory in the calculus of variations, one approach leads to the following function \( f \) : \[ f\left( t\right) \mathrel{\text{:=}} \max \left\{ {{\int }_{0}^{1}L\left( {s, x\left( s\right) ,\dot{x}\left( s\right) }\right) {ds} : \parallel \dot{x}{\parallel }_{2} \leq t}\right\} , \] where the maximum is taken over a certain class of functions \( x : \left\lbrack {0,1}\right\rbrack \rightarrow {\mathbb{R}}^{n} \) , and where the function \( L \) is given. In the presence of the constraint \( \parallel \dot{x}{\parallel }_{2} \leq \) \( t \), the maximum is attained
1012_(GTM178)Nonsmooth Analysis and Control Theory
3
.2, as a consequence of the following implication, which the reader may readily confirm: \[ \zeta \in {\partial }_{P}f\left( x\right) \Rightarrow {Df}\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v. \] While characterizations such as the one given by Theorem 1.4 are of intrinsic interest, it is reassuring to know that they can be and have been of actual use in practice. For example, in developing an existence theory in the calculus of variations, one approach leads to the following function \( f \) : \[ f\left( t\right) \mathrel{\text{:=}} \max \left\{ {{\int }_{0}^{1}L\left( {s, x\left( s\right) ,\dot{x}\left( s\right) }\right) {ds} : \parallel \dot{x}{\parallel }_{2} \leq t}\right\} , \] where the maximum is taken over a certain class of functions \( x : \left\lbrack {0,1}\right\rbrack \rightarrow {\mathbb{R}}^{n} \) , and where the function \( L \) is given. In the presence of the constraint \( \parallel \dot{x}{\parallel }_{2} \leq \) \( t \), the maximum is attained, but the object is to show that the maximum is attained even in the absence of that constraint. The approach hinges upon showing that for \( t \) sufficiently large, the function \( f \) becomes constant. Since \( f \) is increasing by definition, this amounts to showing that \( f \) is (eventually) decreasing, a task that is accomplished in part by Theorem 1.4, since there is no a priori reason for \( f \) to be smooth. This example illustrates how nonsmooth analysis can play a partial but useful role as a tool in the analysis of apparently unrelated issues; detailed examples will be given later in connection with control theory. It is a fact that \( {\partial }_{P}f\left( x\right) \) can in general be empty almost everywhere (a.e.), even when \( f \) is a continuously differentiable function on the real line. Nonetheless, as illustrated by Theorem 1.4, and as we will see in much more complex settings, the proximal subdifferential determines the presence or otherwise of certain basic functional properties. As in the case of the derivative, the utility of \( {\partial }_{P}f \) is based upon the existence of a calculus allowing us to obtain estimates (as in the proximal version of the Mean Value Theorem cited above), or to express the subdifferentials of complicated functionals in terms of the simpler components used to build them. Proximal calculus (among other things) is developed in Chapters 1 and 3, in a Hilbert space setting. ## Generalized Gradients We continue to explore the decrease properties of a given function \( f : {\mathbb{R}}^{n} \rightarrow \) \( \mathbb{R} \), but now we introduce, for the first time, an element of volition: we wish to find a direction in which \( f \) decreases. If \( f \) is smooth, linearization provides an answer: Provided that \( \nabla f\left( x\right) \neq 0 \) , the direction \( v \mathrel{\text{:=}} - \nabla f\left( x\right) \) will do, in the sense that \[ f\left( {x + {tv}}\right) < f\left( x\right) \;\text{ for }t > 0\text{ sufficiently small. } \] (2) What if \( f \) is nondifferentiable? In that case, the proximal subdifferential \( {\partial }_{P}f\left( x\right) \) may not be of any help, as when it is empty, for example. If \( f \) is locally Lipschitz continuous, there is another nonsmooth calculus available, that which is based upon the generalized gradient \( \partial f\left( x\right) \) . A locally Lipschitz function is differentiable almost everywhere; this is Rademacher's Theorem, which is proved in Chapter 3. Its derivative \( {f}^{\prime } \) generates \( \partial f\left( x\right) \) as follows ("co" means "convex hull"): \[ \partial f\left( x\right) = \operatorname{co}\left\{ {\mathop{\lim }\limits_{{i \rightarrow \infty }}\nabla f\left( {x}_{i}\right) : {x}_{i} \rightarrow x,{f}^{\prime }\left( {x}_{i}\right) \text{ exists }}\right\} . \] Then we have the following result on decrease directions: 1.5. Theorem. The generalized gradient \( \partial f\left( x\right) \) is a nonempty compact convex set. If \( 0 \notin \partial f\left( x\right) \), and if \( \zeta \) is the element of \( \partial f\left( x\right) \) having minimal norm, then \( v \mathrel{\text{:=}} - \zeta \) satisfies (2). The calculus of generalized gradients (Chapter 2) will be developed in an arbitrary Banach space, in contrast to proximal calculus. Lest our discussion of decrease become too monotonous, we turn now to another topic, one which will allow us to preview certain geometric concepts that lie at the heart of future developments. For we have learned, since Dini's time, that a better theory results if functions and sets are put on an equal footing. ## 2 Flow-Invariant Sets Let \( S \) be a given closed subset of \( {\mathbb{R}}^{n} \) and let \( \varphi : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{n} \) be locally Lipschitz. The question that concerns us here is whether the trajectories \( x\left( t\right) \) of the differential equation with initial condition \[ \dot{x}\left( t\right) = \varphi \left( {x\left( t\right) }\right) ,\;x\left( 0\right) = {x}_{0}, \] (1) leave \( S \) invariant, in the sense that if \( {x}_{0} \) lies in \( S \), then \( x\left( t\right) \) also belongs to \( S \) for \( t > 0 \) . If this is the case, we say that the system \( \left( {S,\varphi }\right) \) is flow-invariant. As in the previous section (but now for a set rather than a function), linearization provides an answer when the set \( S \) lends itself to it; that is, it is sufficiently smooth. Suppose that \( S \) is a smooth manifold, which means that locally it admits a representation of the form \[ S = \left\{ {x \in {\mathbb{R}}^{n} : h\left( x\right) = 0}\right\} \] where \( h : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{m} \) is a continuously differentiable function with a nonvanishing derivative on \( S \) . Then if the trajectories of (1) remain in \( S \), we have \( h\left( {x\left( t\right) }\right) = 0 \) for \( t \geq 0 \) . Differentiating this for \( t > 0 \) gives \( {h}^{\prime }\left( {x\left( t\right) }\right) \dot{x}\left( t\right) = 0 \) . Substituting \( \dot{x}\left( t\right) = \varphi \left( {x\left( t\right) }\right) \), and letting \( t \) decrease to 0, leads to \[ \left\langle {\nabla {h}_{i}\left( {x}_{0}\right) ,\varphi \left( {x}_{0}\right) }\right\rangle = 0\;\left( {i = 1,2,\ldots, m}\right) . \] The tangent space to the manifold \( S \) at \( {x}_{0} \) is by definition the set \[ \left\{ {v \in {\mathbb{R}}^{n} : \left\langle {\nabla {h}_{i}\left( {x}_{0}\right), v}\right\rangle = 0, i = 1,2,\ldots, m}\right\} , \] so we have proven the necessity part of the following: 2.1. Theorem. Let \( S \) be a smooth manifold. For \( \left( {S,\varphi }\right) \) to be flow-invariant, it is necessary and sufficient that, for every \( x \in S,\varphi \left( x\right) \) belong to the tangent space to \( S \) at \( x \) . There are situations in which we are interested in the flow invariance of a set which is not a smooth manifold, for example, \( S = {\mathbb{R}}_{ + }^{n} \), which corresponds to \( x\left( t\right) \geq 0 \) . It will turn out that it is just as simple to prove the sufficiency part of the above theorem in a nonsmooth setting, once we have decided upon how to define the notion of tangency when \( S \) is an arbitrary closed set. To this end, consider the distance function \( {d}_{S} \) associated with \( S \) : \[ {d}_{S}\left( x\right) \mathrel{\text{:=}} \min \{ \parallel x - s\parallel : s \in S\} \] a globally Lipschitz, nondifferentiable function that turns out to be very useful. Then, if \( x\left( \cdot \right) \) is a solution of (1), where \( {x}_{0} \in S \), we have \( f\left( 0\right) = 0 \) , \( f\left( t\right) \geq 0 \) for \( t \geq 0 \), where \( f \) is the function defined by \[ f\left( t\right) \mathrel{\text{:=}} {d}_{S}\left( {x\left( t\right) }\right) \] What property would ensure that \( f\left( t\right) = 0 \) for \( t \geq 0 \) ; that is, that \( x\left( t\right) \in S \) ? Clearly, that \( f \) be decreasing: monotonicity comes again to the fore! In the light of Theorem 1.1, \( f \) is decreasing iff \( {Df}\left( t\right) \leq 0 \), a condition which at \( t = 0 \) says \[ \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{{d}_{S}\left( {x\left( t\right) }\right) }{t} \leq 0 \] Since \( {d}_{S} \) is Lipschitz, and since we have \[ x\left( t\right) = {x}_{0} + {t\varphi }\left( {x}_{0}\right) + o\left( t\right) \] the lower limit in question is equal to \[ \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{{d}_{S}\left( {{x}_{0} + {t\varphi }\left( {x}_{0}\right) }\right) }{t}. \] This observation suggests the following definition and essentially proves the ensuing theorem, which extends Theorem 2.1 to arbitrary closed sets. 2.2. Definition. A vector \( v \) is tangent to a closed set \( S \) at a point \( x \) if \[ \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{{d}_{S}\left( {x + {tv}}\right) }{t} = 0. \] The set of such vectors is a cone, and is referred to as the Bouligand tangent cone to \( S \) at \( x \), denoted \( {T}_{S}^{B}\left( x\right) \) . It coincides with the tangent space when \( S \) is a smooth manifold. 2.3. Theorem. Let \( S \) be a closed set. Then \( \left( {S,\varphi }\right) \) is flow-invariant iff \[ \varphi \left( x\right) \in {T}_{S}^{B}\left( x\right) \forall x \in S. \] When \( S \) is a smooth manifold, its normal space at \( x \) is defined as the space orthogonal to its tangent space, namely \[ \operatorname{span}\left\{ {\nabla {h}_{i}\left( x\right) : i = 1,2,\ldots, m}\right\} \] and a restatement of Theorem 2.1 in terms of normality goes as follows: \( \left( {S,\varphi }\right) \) is flow-invariant iff \( \langle \zeta ,\varphi \left( x\right) \rangle \leq 0 \) whenever \( x \in S \) and \( \zeta \) is a normal vector to \( S \) at \( x \) . We now consider how to develop in the nonsmooth setting the concept of an outward normal to an arbitrary closed subs
1012_(GTM178)Nonsmooth Analysis and Control Theory
4
uch vectors is a cone, and is referred to as the Bouligand tangent cone to \( S \) at \( x \), denoted \( {T}_{S}^{B}\left( x\right) \) . It coincides with the tangent space when \( S \) is a smooth manifold. 2.3. Theorem. Let \( S \) be a closed set. Then \( \left( {S,\varphi }\right) \) is flow-invariant iff \[ \varphi \left( x\right) \in {T}_{S}^{B}\left( x\right) \forall x \in S. \] When \( S \) is a smooth manifold, its normal space at \( x \) is defined as the space orthogonal to its tangent space, namely \[ \operatorname{span}\left\{ {\nabla {h}_{i}\left( x\right) : i = 1,2,\ldots, m}\right\} \] and a restatement of Theorem 2.1 in terms of normality goes as follows: \( \left( {S,\varphi }\right) \) is flow-invariant iff \( \langle \zeta ,\varphi \left( x\right) \rangle \leq 0 \) whenever \( x \in S \) and \( \zeta \) is a normal vector to \( S \) at \( x \) . We now consider how to develop in the nonsmooth setting the concept of an outward normal to an arbitrary closed subset \( S \) of \( {\mathbb{R}}^{n} \) . The key is projection: Given a point \( u \) not in \( S \), and let \( x \) be a point in \( S \) that is closest to \( u \) ; we say that \( x \) lies in the projection of \( u \) onto \( S \) . Then the vector \( u - x \) (and all its nonnegative multiples) defines a proximal normal direction to \( S \) at \( x \) . The set of all vectors constructed this way (for fixed \( x \), by varying \( u) \) is called the proximal normal cone to \( S \) at \( x \), and denoted \( {N}_{S}^{P}\left( x\right) \) . It coincides with the normal space when \( S \) is a smooth manifold. It is possible to characterize flow-invariance in terms of proximal normals as follows: 2.4. Theorem. Let \( S \) be a closed set. Then \( \left( {S,\varphi }\right) \) is flow-invariant iff \[ \langle \zeta ,\varphi \left( x\right) \rangle \leq 0\forall \zeta \in {N}_{S}^{P}\left( x\right) ,\forall x \in S. \] We can observe a certain duality between Theorems 2.3 and 2.4. The former characterizes flow-invariance in terms internal to the set \( S \), via tangency, while the latter speaks of normals generated by looking outside the set. In the case of a smooth manifold, the duality is exact: the tangential and normal conditions are restatements of one another. In the general non-smooth case, this is no longer true (pointwise, the sets \( {T}_{S}^{B} \) and \( {N}_{S}^{P} \) are not obtainable one from the other). While the word "duality" may have to be interpreted somewhat loosely, this element is an important one in our overall approach to developing non-smooth analysis. The dual objects often work well in tandem. For example, while tangents are often convenient to verify flow-invariance, proximal normals lie at the heart of the "proximal aiming method" used in Chapter 4 to define stabilizing feedbacks. Another type of duality that we seek involves coherence between the various analytical and geometrical constructs that we define. To illustrate this, consider yet another approach to studying the flow-invariance of \( \left( {S,\varphi }\right) \), that which seeks to characterize the property (cited above) that the function \( f\left( t\right) = {d}_{S}\left( {x\left( t\right) }\right) \) be decreasing in terms of the proximal subdifferential of \( f \) (rather than subderivates). If an appropriate "chain rule" is available, then we could hope to use it in conjunction with Theorem 1.4 in order to reduce the question to an inequality: \[ \left\langle {{\partial }_{P}{d}_{S}\left( x\right) ,\varphi \left( x\right) }\right\rangle \leq 0\forall x \in S. \] Modulo some technicalities that will interest us later, this is feasible. In the light of Theorem 2.4, we are led to suspect (or hope for) the following fact: \[ {N}_{S}^{P}\left( x\right) = \text{the cone generated by}{\partial }_{P}{d}_{S}\left( x\right) \text{.} \] This type of formula illustrates what we mean by coherence between constructs, in this case between the proximal normal cone to a set and the proximal subdifferential of its distance function. ## 3 Optimization As a first illustration of how nonsmoothness arises in the subject of optimization, we consider minimax problems. Let a smooth function \( f \) depend on two variables \( x \) and \( u \), where the first is thought of as being a choice variable, while the second cannot be specified; it is known only that \( u \) varies in a set \( M \) . We seek to minimize \( f \) . Corresponding to a choice of \( x \), the worst possibility over the values of \( u \) that may occur corresponds to the following value of \( f : \mathop{\max }\limits_{{u \in M}}f\left( {x, u}\right) \) . Accordingly, we consider the problem \[ \mathop{\operatorname{minimize}}\limits_{x}g\left( x\right) ,\text{ where }g\left( x\right) \mathrel{\text{:=}} \mathop{\max }\limits_{{u \in M}}f\left( {x, u}\right) . \] The function \( g \) so defined will not generally be smooth, even if \( f \) is a nice function and the maximum defining \( g \) is attained. To see this in a simple setting, consider the upper envelope \( g \) of two smooth functions \( {f}_{1},{f}_{2} \) . (We suggest that the reader make a sketch at this point.) Then \( g \) will have a corner at a point \( x \) where \( {f}_{1}\left( x\right) = {f}_{2}\left( x\right) \), provided that \[ {f}_{1}^{\prime }\left( x\right) \neq {f}_{2}^{\prime }\left( x\right) \] Returning to the general case, we remark that under mild hypotheses, the generalized gradient \( \partial g\left( x\right) \) can be calculated; we find \[ \partial g\left( x\right) = \operatorname{co}\left\{ {{f}_{x}^{\prime }\left( {x, u}\right) : u \in M\left( x\right) }\right\} \] where \[ M\left( x\right) \mathrel{\text{:=}} \{ u \in M : f\left( {x, u}\right) = g\left( x\right) \} . \] This characterization can then serve as the initial step in approaching the problem, either analytically or numerically. There may then be explicit constraints on \( x \) to consider. A problem having a very specific structure, and one which is of considerable importance in engineering and optimal design, is the following eigenvalue problem. Let the \( n \times n \) symmetric matrix \( A \) depend on a parameter \( x \) in some way, so that we write \( A\left( x\right) \) . A familiar criterion in designing the underlying system which is represented by \( A\left( x\right) \) is that the maximal eigenvalue \( \Lambda \) of \( A\left( x\right) \) be made as small as possible. This could correspond to a question of stability, for example. It turns out that this problem is of minimax type, for by Rayleigh's formula for the maximal eigenvalue we have \[ \Lambda \left( x\right) = \max \{ \langle u, A\left( x\right) u\rangle : \parallel u\parallel = 1\} . \] The function \( \Lambda \left( \cdot \right) \) will generally be nonsmooth, even if the dependence \( x \mapsto A\left( x\right) \) is itself smooth. For example, the reader may verify that the maximal eigenvalue \( \Lambda \left( {x, y}\right) \) of the matrix \[ A\left( {x, y}\right) \mathrel{\text{:=}} \left\lbrack \begin{matrix} 1 + x & y \\ y & 1 - x \end{matrix}\right\rbrack \] is given by \( 1 + \parallel \left( {x, y}\right) \parallel \) . Note that the minimum of this function occurs at \( \left( {0,0}\right) \), precisely its point of nondifferentiability. This is not a coincidence, and it is now understood that nondifferentiability is to be expected as an intrinsic feature of design problems generally, in problems as varied as designing an optimal control or finding the shape of the strongest column. Another class of problems in which nondifferentiability plays a role is that of \( {L}^{1} \) -optimization. In its discrete version, the problem consists of minimizing a function \( f \) of the form \[ f\left( x\right) \mathrel{\text{:=}} \mathop{\sum }\limits_{{i = 1}}^{p}{m}_{i}\begin{Vmatrix}{x - {s}_{i}}\end{Vmatrix}. \] (1) Such problems arise, for example, in approximation and statistics, where \( {L}^{1} \) -approximation possesses certain features that can make it preferable to the more familiar (and smooth) \( {L}^{2} \) -approximation. Let us examine such a problem in the context of a simple physical system. ## Torricelli's Table A table has holes in it at points whose coordinates are \( {s}_{1},{s}_{2},\ldots ,{s}_{p} \) . Strings are attached to masses \( {m}_{1},{m}_{2},\ldots ,{m}_{p} \), passed through the corresponding hole, and then are all tied to a point mass \( m \) whose position is denoted \( x \) (see Figure 0.1). If friction and the weight of the strings are negligible, the equilibrium position \( x \) of the nexus is precisely the one that minimizes the function \( f \) given by (1), since \( f\left( x\right) \) can be recognized as the potential energy of the system. The proximal subdifferential of the function \( x \mapsto \parallel x - s\parallel \) is the closed unit ball if \( x = s \), and otherwise is the singleton set consisting of its derivative, the point \( \left( {x - s}\right) /\parallel x - s\parallel \) . Using this fact, and some further calculus, we can derive the following necessary condition for a point \( x \) to minimize \( f \) ; \[ 0 \in \mathop{\sum }\limits_{{i = 1}}^{p}{m}_{i}{\partial }_{P}\parallel \left( \cdot \right) - {s}_{i}\parallel \left( x\right) \] (2) ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_25_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_25_0.jpg) FIGURE 0.1. Torricelli's table. Of course, (2) is simply Fermat's rule in subdifferential terms, interpreted for the particular function \( f \) that we are dealing with. There is not necessarily a unique point \( x \) that satisfies relation (2), but it is the case that any point satisfying (2) globally minimizes \( f \) . This is because \( f \) is convex, another functional class that plays an important role in the subject. A consequence of convexity is that there are no purely local minima in this problem. When \( p =
1012_(GTM178)Nonsmooth Analysis and Control Theory
5
g of its derivative, the point \( \left( {x - s}\right) /\parallel x - s\parallel \) . Using this fact, and some further calculus, we can derive the following necessary condition for a point \( x \) to minimize \( f \) ; \[ 0 \in \mathop{\sum }\limits_{{i = 1}}^{p}{m}_{i}{\partial }_{P}\parallel \left( \cdot \right) - {s}_{i}\parallel \left( x\right) \] (2) ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_25_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_25_0.jpg) FIGURE 0.1. Torricelli's table. Of course, (2) is simply Fermat's rule in subdifferential terms, interpreted for the particular function \( f \) that we are dealing with. There is not necessarily a unique point \( x \) that satisfies relation (2), but it is the case that any point satisfying (2) globally minimizes \( f \) . This is because \( f \) is convex, another functional class that plays an important role in the subject. A consequence of convexity is that there are no purely local minima in this problem. When \( p = 3 \), each \( {m}_{i} = 1 \), and the three points are the vertices of a triangle, the problem becomes that of finding a point such that the sum of its distances from the vertices is minimal. The solution is called the Torricelli point, after the seventeenth-century mathematician. The fact that (2) is necessary and sufficient for a minimum allows us to recover easily certain classical conclusions regarding this problem. As an example, the reader is invited to establish that the Torricelli point coincides with a vertex of the triangle iff the angle at that vertex is \( {120}^{ \circ } \) or more. Returning now to the general case of our table, it is possible to make the system far more complex by the addition of one more string and one more mass \( {m}_{0} \), if we allow that mass to hang over the outside edge of the table. Then the extra string will automatically trace a line segment from \( x \) to a point \( s\left( x\right) \) on the edge of the table that is closest to \( x \) (locally at least, in the sense that \( s\left( x\right) \) is the closest point to \( x \) on the edge, relative to a neighborhood of \( s\left( x\right) \) .) If \( S \) is the set defined as the closure of the complement of the table, the potential energy (up to a constant) of the ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_26_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_26_0.jpg) FIGURE 0.2. Discontinuity of the local projection. system is now, at its lowest level, \[ \widetilde{f}\left( x\right) \mathrel{\text{:=}} {m}_{0}{d}_{S}\left( x\right) + \mathop{\sum }\limits_{{i = 1}}^{p}{m}_{i}\begin{Vmatrix}{x - {s}_{i}}\end{Vmatrix}. \] The function \( \widetilde{f} \) is not only nonsmooth, as was \( f \), but also nonconvex, and will admit local minima at different energy levels. The points \( s \) on the boundary of \( S \) which are feasible as points through which would pass the over-the-table string (at equilibrium) are precisely those for which the proximal normal cone \( {N}_{S}^{P}\left( s\right) \) is nonzero. Such points can be rather sparse, though they are always dense in the boundary of \( S \) . For a rectangular table, there are exactly four points at which \( {N}_{S}^{P} \) is \( \{ 0\} \) . If \( x\left( t\right) \) represents a displacement undergone by the nexus over time, Newton's Law implies \[ M\ddot{x} = \mathop{\sum }\limits_{{i = 1}}^{p}{m}_{i}\frac{\left( {s}_{i} - x\right) }{\begin{Vmatrix}{s}_{i} - x\end{Vmatrix}} + {m}_{0}\frac{s\left( x\right) - x}{\parallel s\left( x\right) - x\parallel } \] (3) on any time interval during which \( x \neq {s}_{i}, x \neq s\left( x\right) \), where \( M \) is the total mass of the system, namely \( m + {m}_{0} + \sum {m}_{i} \) . The local projection \( x \mapsto s\left( x\right) \) will be discontinuous in general, so in solving (3), there arises the issue of a differential equation incorporating a discontinuous function of the state. Figure 0.2 illustrates the discontinuity of \( s\left( x\right) \) in a particular case. As \( x \) traces the line segment from \( u \) toward \( v \), the corresponding \( s\left( x\right) \) traces the segment joining \( A \) and \( B \) . When \( x \) goes beyond \( v, s\left( x\right) \) abruptly moves to the vicinity of the point \( C \) . (The figure omits all the strings acting upon \( x \) .) We will treat the issue of discontinuous differential equations in Chapter 4, where it arises in connection with feedback control design. ## Constrained Optimization In minimizing a function \( f\left( x\right) \), it is often necessary to take account of explicit constraints on the point \( x \), for example, that \( x \) lie in a given set \( S \) . There are two methods for dealing with such problems that figure prominently in this book. The first of these, called exact penalization, seeks to replace the constrained optimization problem \[ \text{minimize}f\left( x\right) \text{subject to}x \in S \] by the unconstrained problem \[ \text{minimize}f\left( x\right) + K{d}_{S}\left( x\right) \text{,} \] where \( {d}_{S} \) is the distance function introduced in \( §2 \) . Under mild conditions, this constraint-removal technique is justified, for \( K \) sufficiently large. Since the distance function is never differentiable at all boundary points of \( S \), however, and since that is precisely where the solutions of the new problem are likely to lie, we are subsequently obliged to deal with a nonsmooth minimization problem, even if the original problem has smooth data \( f, S \) . The second general technique for dealing with constrained optimization, called value function analysis, is applied when the constraint set \( S \) has an explicit functional representation, notably in terms of equalities and inequalities. A simple case to illustrate: we seek to minimize \( f\left( x\right) \) subject to \( h\left( x\right) = 0 \) . Let us embed the problem in a family of similar ones, parametrized by a perturbation term in the equality constraint. Specifically, the problem \( P\left( \alpha \right) \) is the following: \[ P\left( \alpha \right) \text{: minimize}f\left( x\right) \text{over}x\text{subject to}h\left( x\right) + \alpha = 0\text{.} \] Let \( V\left( \alpha \right) \), the associated value function of this perturbation scheme, designate the minimum value of the problem \( P\left( \alpha \right) \) . Our original problem is simply \( P\left( 0\right) \) . If \( {x}_{0} \) is a solution of \( P\left( 0\right) \), then of course \( h\left( {x}_{0}\right) = 0 \) (since \( {x}_{0} \) must be feasible for \( P\left( 0\right) \) ), and we have \( V\left( 0\right) = \) \( f\left( {x}_{0}\right) \) . This last observation implies that \[ f\left( {x}_{0}\right) - V\left( {-h\left( {x}_{0}\right) }\right) = 0, \] whereas it follows from the very definition of \( V \) that, for any \( x \) whatsoever, we have \[ f\left( x\right) - V\left( {-h\left( x\right) }\right) \geq 0. \] (We ask our readers to convince themselves of this.) Put another way, these observations amount to saying that the function \[ x \mapsto f\left( x\right) - V\left( {-h\left( x\right) }\right) \] attains a minimum at \( x = {x}_{0} \), whence \[ {f}^{\prime }\left( {x}_{0}\right) + {V}^{\prime }\left( 0\right) {h}^{\prime }\left( {x}_{0}\right) = 0 \] a conclusion that we recognize as the Lagrange Multiplier Rule (with, as a bonus, a sensitivity interpretation of the multiplier, \( {V}^{\prime }\left( 0\right) \) ). If our readers are dubious about this simple proof of the Multiplier Rule, they are justified in being so. Still, the only fallacy involved is the implicit assumption that \( V \) is differentiable. Nonsmooth analysis will allow us to develop a rigorous argument along the lines of the above, in Chapter 3. ## 4 Control Theory In the control theory of ordinary differential equations, the standard model revolves around the system \[ \dot{x}\left( t\right) = f\left( {x\left( t\right), u\left( t\right) }\right) \text{ a.e.,}\;0 \leq t \leq T, \] (1) where the (measurable) control function \( u\left( \cdot \right) \) is chosen subject to the constraint \[ u\left( t\right) \in U\text{ a.e.,} \] (2) and where the ensuing state \( x\left( \cdot \right) \) is subject to an initial condition \( x\left( 0\right) = {x}_{0} \) and perhaps other constraints. This indirect control of \( x\left( \cdot \right) \) via the choice of \( u\left( \cdot \right) \) is to be exercised for a purpose, of which there are two principal sorts: positional \( \left( {x\left( t\right) }\right. \) is to remain in a given set in \( {\mathbb{R}}^{n} \), or approach that set) and optimal \( \left( {x\left( \cdot \right) \text{, together with}u\left( \cdot \right) \text{, is to minimize a given functional}}\right) \) . As is the case in optimization, certain problems arise in which the underlying data are nonsmooth; minimax criteria are an example. In this section, however, we wish to convey to the reader how considerations of nondiffer-entiability arise from the very way in which we might hope to solve the problem. Our illustrative example will be one that combines positional and optimal considerations, namely the minimal time problem. It consists of finding the least \( T \geq 0 \) together with a control function \( u\left( \cdot \right) \) on \( \left\lbrack {0, T}\right\rbrack \) having the property that the resulting state \( x \) satisfies \( x\left( T\right) = 0 \) . Informally, it is required to steer the initial state \( {x}_{0} \) to the origin in least time. Let us introduce the following set-valued mapping \( F \) : \[ F\left( x\right) \mathrel{\text{:=}} f\left( {x, U}\right) . \] A trajectory of \( F \) on an interval \( \left\lbrack {0, T}\right\rbrack \) is an absolutely continuous function \( x\left( \cdot \right) \) on \( \left\lbrack {0, T}\right\rbrack \) which satisfies \[ \dot{x}\left( t\right) \in F\left( {x\left( t\right) }\right) \text{ a.e.,}\;0
1012_(GTM178)Nonsmooth Analysis and Control Theory
6
e reader how considerations of nondiffer-entiability arise from the very way in which we might hope to solve the problem. Our illustrative example will be one that combines positional and optimal considerations, namely the minimal time problem. It consists of finding the least \( T \geq 0 \) together with a control function \( u\left( \cdot \right) \) on \( \left\lbrack {0, T}\right\rbrack \) having the property that the resulting state \( x \) satisfies \( x\left( T\right) = 0 \) . Informally, it is required to steer the initial state \( {x}_{0} \) to the origin in least time. Let us introduce the following set-valued mapping \( F \) : \[ F\left( x\right) \mathrel{\text{:=}} f\left( {x, U}\right) . \] A trajectory of \( F \) on an interval \( \left\lbrack {0, T}\right\rbrack \) is an absolutely continuous function \( x\left( \cdot \right) \) on \( \left\lbrack {0, T}\right\rbrack \) which satisfies \[ \dot{x}\left( t\right) \in F\left( {x\left( t\right) }\right) \text{ a.e.,}\;0 \leq t \leq T. \] (3) Under mild hypotheses, it is a fact that \( x\left( \cdot \right) \) is a trajectory (i.e., satisfies (3)) iff there is a control function \( u\left( \cdot \right) \) (i.e., a measurable function \( u\left( \cdot \right) \) satisfying (2)) for which the differential equation (1) linking \( x \) and \( u \) holds. (See Chapter 3 for this; here, we are not even going to state hypotheses at all.) In terms of trajectories, then, the problem is to find one which is optimal from \( {x}_{0} \) ; that is, one which reaches the origin as quickly as possible. Let us undertake the quest. We begin by introducing the minimal time function \( T\left( \cdot \right) \), defined on \( {\mathbb{R}}^{n} \) as follows: \[ T\left( \alpha \right) \mathrel{\text{:=}} \min \{ T \geq 0 : \text{ some trajectory }x\left( \cdot \right) \text{ satisfies }x\left( 0\right) = \alpha, x\left( T\right) = 0.\} \] An issue of controllability arises here: Is it always possible to steer \( \alpha \) to 0 in finite time? We will study this question in Chapter 4; for now, let us assume that such is the case. The principle of optimality is the dual observation that if \( x\left( \cdot \right) \) is any trajectory, the function \[ t \mapsto T\left( {x\left( t\right) }\right) + t \] is increasing, and that if \( x \) is optimal, then the same function is constant. In other terms, if \( x\left( \cdot \right) \) is an optimal trajectory joining \( \alpha \) to 0, then \[ T\left( {x\left( t\right) }\right) = T\left( \alpha \right) - t\text{ for }0 \leq t \leq T\left( \alpha \right) , \] since an optimal trajectory from the point \( x\left( t\right) \) is furnished by the truncation of \( x\left( \cdot \right) \) to the interval \( \left\lbrack {t, T\left( \alpha \right) }\right\rbrack \) . If \( x\left( \cdot \right) \) is any trajectory, then the inequality \[ T\left( {x\left( t\right) }\right) \geq T\left( \alpha \right) - t \] is a reflection of the fact that in going to the point \( x\left( t\right) \) from \( \alpha \) (in time \( t \) ), we may have acted optimally (in which case equality holds) or not (then inequality holds). Since \( t \mapsto T\left( {x\left( t\right) }\right) + t \) is increasing, we expect to have \[ \langle \nabla T\left( {x\left( t\right) }\right) ,\dot{x}\left( t\right) \rangle + 1 \geq 0 \] (4) with equality when \( x\left( \cdot \right) \) is an optimal trajectory. The possible values of \( \dot{x}\left( t\right) \) for a trajectory being precisely the elements of the set \( F\left( {x\left( t\right) }\right) \), we arrive at \[ \mathop{\min }\limits_{{v \in F\left( x\right) }}\{ \langle \nabla T\left( x\right), v\rangle \} + 1 = 0. \] We define the (lower) Hamiltonian function \( h \) as follows: \[ h\left( {x, p}\right) \mathrel{\text{:=}} \mathop{\min }\limits_{{v \in F\left( x\right) }}\langle p, v\rangle . \] In terms of \( h \), the partial differential equation obtained above reads \[ h\left( {x,\nabla T\left( x\right) }\right) + 1 = 0, \] (5) a special case of the Hamilton-Jacobi equation. Here is the first step in our quest: use the Hamilton-Jacobi equation (5), together with the boundary condition \( T\left( 0\right) = 0 \), to find \( T\left( \cdot \right) \) . How will this help us find the optimal trajectory? To answer this question, we recall that an optimal trajectory is such that equality holds in (4). This suggests the following procedure: for each \( x \), let \( \widehat{v}\left( x\right) \) be a point in \( F\left( x\right) \) satisfying \[ \mathop{\min }\limits_{{v \in F\left( x\right) }}\langle \nabla T\left( x\right), v\rangle = \langle \nabla T\left( x\right) ,\widehat{v}\left( x\right) \rangle = - 1. \] (6) Then, if we construct \( x\left( \cdot \right) \) via the initial-value problem \[ \dot{x}\left( t\right) = \widehat{v}\left( {x\left( t\right) }\right) ,\;x\left( 0\right) = \alpha , \] (7) we will have a trajectory that is optimal (from \( \alpha \) )! Here is why: Let \( x\left( \cdot \right) \) satisfy (7); then \( x\left( \cdot \right) \) is a trajectory, since \( \widehat{v}\left( x\right) \) belongs to \( F\left( x\right) \) . Furthermore, \[ \frac{d}{dt}T\left( {x\left( t\right) }\right) = \langle \nabla T\left( {x\left( t\right) }\right) ,\dot{x}\left( t\right) \rangle \] \[ = \langle \nabla T\left( {x\left( t\right) }\right) ,\widehat{v}\left( {x\left( t\right) }\right) \rangle = - 1. \] In consequence, we find \[ T\left( {x\left( t\right) }\right) = T\left( \alpha \right) - t \] which implies that at \( t = T\left( \alpha \right) \), we must have \( x = 0 \) . Therefore \( x\left( \cdot \right) \) is an optimal trajectory. Let us stress the important point that \( \widehat{v}\left( \cdot \right) \) generates the optimal trajectory from any initial value \( \alpha \) (via (7)), and so constitutes what can be considered the Holy Grail for this problem: an optimal feedback synthesis. There can be no more satisfying answer to the problem: If you find yourself at \( x \), just choose \( \dot{x} = \widehat{v}\left( x\right) \) to approach the origin as fast as possible. Unfortunately, there are serious obstacles to following the route that we have just outlined, beginning with the fact that \( T \) is nondifferentiable, as simple examples show. ( \( T \) is a value function, analogous to the one we met in \( §3 \) .) We will therefore have to examine anew the argument that led to the Hamilton-Jacobi equation (5), which in any case, will have to be recast in some way to accommodate nonsmooth solutions. Having done so, will the generalized Hamilton-Jacobi equation admit \( T \) as the unique solution? The next step (after characterizing \( T \) ) offers fresh difficulties of its own. Even if \( T \) were smooth, there would be in general no continuous function \( \widehat{v}\left( \cdot \right) \) satisfying (6) for each \( x \) . The meaning and existence of a trajectory \( x\left( \cdot \right) \) generated by \( \widehat{v}\left( \cdot \right) \) via (7) is therefore problematic in itself. The intrinsic difficulties of the "dynamic programming" approach to the minimal-time problem, which is what we have outlined above, have been an historical focal point of activity in differential equations and control, and it is only recently that fully satisfying answers to all the questions raised above have been found. We will present them in Chapter 4, together with results bearing on other basic topics in mathematical control theory: invariance, equilibria, stability, and necessary and sufficient conditions for optimality. Let us begin now to be more precise. ## 5 Notation We expect our readers to have taken a course in functional analysis, and we hope that the following notation appears natural to them. \( X \) is a real Hilbert space or Banach space with norm \( \parallel \cdot \parallel \) . The open ball in \( X \) (of radius 1, centered at 0) is denoted by \( B \), its closure by \( \bar{B} \) . We also write \( {B}_{X} \) if \( X \) is to be distinguished from other spaces. The inner product of \( \zeta \) and \( x \) is denoted \( \langle \zeta, x\rangle \), a notation which is also employed when \( X \) is a Banach space for the evaluation, at \( x \in X \), of the linear functional \( \zeta \in {X}^{ * } \) (the space of continuous linear functionals defined on \( X \) ). The open unit ball in \( {X}^{ * } \) is written \( {B}_{ * } \) . The notation \[ x = \underset{i \rightarrow \infty }{\text{ lim }}{x}_{i} \] means that the sequence \( \left\{ {x}_{i}\right\} \) converges weakly to \( x \) . Similarly, \( {w}^{ * } \) refers to the weak* topology on the space \( {X}^{ * } \cdot {L}_{n}^{p}\left\lbrack {a, b}\right\rbrack \) refers to the set of \( p \) -integrable functions from \( \left\lbrack {a, b}\right\rbrack \) to \( {\mathbb{R}}^{n} \) . For the two subsets \( {S}_{1} \) and \( {S}_{2} \) of \( X \), the set \( {S}_{1} + {S}_{2} \) is given by \[ \left\{ {s = {s}_{1} + {s}_{2} : {s}_{1} \in {S}_{1},{s}_{2} \in {S}_{2}}\right\} . \] The open ball of radius \( r > 0 \), centered at \( x \), is denoted by either \( B\left( {x;r}\right) \) or \( x + {rB} \), where (strictly speaking) the latter should be written \( \{ x\} + {rB} \) . The closure of \( B\left( {x;r}\right) \) is written as either \( \bar{B}\left( {x;r}\right) \) or \( x + r\bar{B} \) . We confess to writing "iff" for "if and only if." The symbol \( \mathrel{\text{:=}} \) means "equal by definition." We lean toward mnemonic notation in general. For a given set \( S \), the expressions \[ \text{int}S,\;\operatorname{cl}S,\;\operatorname{bdry}S,\;\operatorname{co}S,\;\overline{\operatorname{co}}S\text{,} \] signify the interior, closure, boundary, convex hull, and closed convex hull of \( S \), respectively. A list of the principal notational constructs used in the book is given in the Notes and Comments at the end. A reference such as Theorem 1.2.3 refers
1012_(GTM178)Nonsmooth Analysis and Control Theory
7
y \[ \left\{ {s = {s}_{1} + {s}_{2} : {s}_{1} \in {S}_{1},{s}_{2} \in {S}_{2}}\right\} . \] The open ball of radius \( r > 0 \), centered at \( x \), is denoted by either \( B\left( {x;r}\right) \) or \( x + {rB} \), where (strictly speaking) the latter should be written \( \{ x\} + {rB} \) . The closure of \( B\left( {x;r}\right) \) is written as either \( \bar{B}\left( {x;r}\right) \) or \( x + r\bar{B} \) . We confess to writing "iff" for "if and only if." The symbol \( \mathrel{\text{:=}} \) means "equal by definition." We lean toward mnemonic notation in general. For a given set \( S \), the expressions \[ \text{int}S,\;\operatorname{cl}S,\;\operatorname{bdry}S,\;\operatorname{co}S,\;\overline{\operatorname{co}}S\text{,} \] signify the interior, closure, boundary, convex hull, and closed convex hull of \( S \), respectively. A list of the principal notational constructs used in the book is given in the Notes and Comments at the end. A reference such as Theorem 1.2.3 refers to Theorem 2.3 of Chapter 1, which will be found in §2.3. From within Chapter 1, it is referred to simply as Theorem 2.3. 1 Proximal Calculus in Hilbert Space -Dangerous Liaisons, the Film. We introduce in this chapter two basic constructs of nonsmooth analysis: proximal normals (to a set) and proximal subgradients (of a function). Proximal normals are direction vectors pointing outward from a set, generated by projecting a point onto the set. Proximal subgradients have a certain local support property to the epigraph of a function. It is a familiar device to view a function as a set (through its graph), but we develop the duality between functions and sets to a much greater extent, extending it to include the calculus of these normals and subgradients. The very existence of a proximal subgradient often says something of interest about a function at a point; the Density Theorem of \( §3 \) is a deep result affirming existence on a substantial set. From it we deduce two minimization principles. These are theorems bearing upon situations where a minimum is "almost attained," and which assert that a small perturbation leads to actual attainment. We will meet some useful classes of functions along the way: convex, Lipschitz, indicator, and distance functions. Finally, we will see some elements of proximal calculus, notably the sum and chain rules. ## 1 Closest Points and Proximal Normals Let \( X \) be a real Hilbert space, and let \( S \) be a nonempty subset of \( X \) . Suppose that \( x \) is a point not lying in \( S \) . Suppose further that there exists ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_34_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_34_0.jpg) FIGURE 1.1. A set \( S \) and some of its boundary points. a point \( s \) in \( S \) whose distance to \( x \) is minimal. Then \( s \) is called a closest point or a projection of \( x \) onto \( S \) . The set of all such closest points is denoted by \( {\operatorname{proj}}_{S}\left( x\right) \) . It is clear that \( s \in {\operatorname{proj}}_{S}\left( x\right) \) iff \( \{ s\} \subset S \cap \bar{B}\left( {x;\parallel x - s\parallel }\right) \) and \( S \cap B\left( {x;\parallel x - s\parallel }\right) = \varnothing \) . See Figure 1.1. The vector \( x - s \) determines what we will call a proximal normal direction to \( S \) at \( s \) ; any nonnegative multiple \( \zeta = t\left( {x - s}\right), t \geq 0 \), of such a vector will be called a proximal normal (or a \( P \) -normal) to \( S \) at \( s \) . The set of all \( \zeta \) obtainable in this manner is termed the proximal normal cone to \( S \) at \( s \), and is denoted by \( {N}_{S}^{P}\left( s\right) \) ; it is clear that \( {N}_{S}^{P}\left( s\right) \) is in fact a cone; i.e., a set closed under forming nonnegative scalar multiples. Intuitively, a proximal normal vector at a given point defines a direction of perpendicular departure from the set. Suppose \( s \in S \) is such that \( s \notin {\operatorname{proj}}_{S}\left( x\right) \) for all \( x \) not in \( S \) (which is certainly the case if \( s \) lies in int \( S \) ). Then we set \( {N}_{S}^{P}\left( s\right) = \{ 0\} \) . When \( s \notin S \) , then \( {N}_{S}^{P}\left( s\right) \) remains undefined. In Figure 1.1, the points \( {s}_{3} \) and \( {s}_{5} \) have \( P \) -normal cones equal to \( \{ 0\} \), and the points \( {s}_{1},{s}_{2},{s}_{7} \), and \( {s}_{8} \) have at least two independent vectors in their \( P \) -normal cones. The remaining boundary points of \( S \) have their \( P \) -normal cone generated by a single nonzero vector. Notice that we have not asserted above that the point \( x \) must admit a closest point \( s \) in \( S \) . In finite dimensions, there is little difficulty in assuring that projections exist, for it suffices that \( S \) be closed. We will in fact only focus on closed sets \( S \), but nonetheless, the issue of the existence of closest points in infinite dimensions is far more subtle, and will be an important point later. 1.1. Exercise. Let \( X \) admit a countable orthonormal basis \( {\left\{ {e}_{i}\right\} }_{i = 1}^{\infty } \) , and set \[ S \mathrel{\text{:=}} \left\{ {\frac{i + 1}{i}{e}_{i} : i \geq 1}\right\} \] Prove that \( S \) is closed, and that \( {\operatorname{proj}}_{S}\left( 0\right) = \varnothing \) . The above concepts can be described in terms of the distance function \( {d}_{S} : X \rightarrow \mathbb{R} \), which is given by \[ {d}_{S}\left( x\right) \mathrel{\text{:=}} \inf \{ \parallel x - s\parallel : s \in S\} . \] Occasionally it is more convenient to write \( d\left( {x;S}\right) \) for \( {d}_{S}\left( x\right) \) . The set \( {\operatorname{proj}}_{S}\left( x\right) \) consists of those points (if any) at which the infimum defining \( {d}_{S}\left( x\right) \) is attained. We also have the formula \[ {N}_{S}^{P}\left( s\right) = \left\{ {\zeta : \exists t > 0\text{ so that }{d}_{S}\left( {s + {t\zeta }}\right) = t\parallel \zeta \parallel }\right\} . \] Some further basic properties of \( {d}_{S} \) are listed in the following exercise: 1.2. Exercise. (a) Show that \( x \) belongs to \( \operatorname{cl}S \) iff \( {d}_{S}\left( x\right) = 0 \) . (b) Suppose that \( S \) and \( {S}^{\prime } \) are two subsets of \( X \) . Show that \( {d}_{S} = \) \( {d}_{{S}^{\prime }} \) iff \( \operatorname{cl}S = \operatorname{cl}{S}^{\prime } \) . (c) Show that \( {d}_{S} \) satisfies \[ \left| {{d}_{S}\left( x\right) - {d}_{S}\left( y\right) }\right| \leq \parallel x - y\parallel \forall x, y \in X \] which says that \( {d}_{S} \) is Lipschitz of rank 1, on \( X \) . (d) If \( S \) is a closed subset of \( {\mathbb{R}}^{n} \), show that \( {\operatorname{proj}}_{S}\left( x\right) \neq \varnothing \) for all \( x \) , and that the set \( \left\{ {s \in {\operatorname{proj}}_{S}\left( x\right) : x \in {\mathbb{R}}^{n} \smallsetminus S}\right\} \) is dense in bdry \( S \) . (Hint. Let \( s \in \operatorname{bdry}S \), and let \( \left\{ {x}_{i}\right\} \) be a sequence not in \( S \) that converges to \( s \) . Show that any sequence \( \left\{ {s}_{i}\right\} \) chosen with \( \left. {{s}_{i} \in \operatorname{proj}{x}_{i}\text{, converges to }s\text{.}}\right) \) Suppose now that \( s \in {\operatorname{proj}}_{S}\left( x\right) \) . This is equivalent to the condition \[ \begin{Vmatrix}{x - {s}^{\prime }}\end{Vmatrix} \geq \parallel x - s\parallel \forall {s}^{\prime } \in S. \] If we square both sides of this inequality and expand in terms of the inner product, we thus obtain the conclusion that \( s \in {\operatorname{proj}}_{S}\left( x\right) \) iff \[ \left\langle {x - s,{s}^{\prime } - s}\right\rangle \leq \frac{1}{2}{\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2}\forall {s}^{\prime } \in S. \] This in turn is clearly equivalent to \[ \left\langle {\left\lbrack {s + t\left( {x - s}\right) }\right\rbrack - s,{s}^{\prime } - s}\right\rangle \leq \frac{1}{2}{\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2}\forall t \in \left\lbrack {0,1}\right\rbrack ,\forall {s}^{\prime } \in S, \] ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_36_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_36_0.jpg) FIGURE 1.2. A point \( {x}_{1} \) and its five projections. which (by the preceding characterization) holds iff for all \( t \in \left\lbrack {0,1}\right\rbrack \), we have \( s \in {\operatorname{proj}}_{S}\left( {s + t\left( {x - s}\right) }\right) \) . These remarks are summarized in the following: 1.3. Proposition. Let \( S \) be a nonempty subset of \( X \), and let \( x \in X, s \in S \) . The following are equivalent: (a) \( s \in {\operatorname{proj}}_{S}\left( x\right) \) ; (b) \( s \in {\operatorname{proj}}_{S}\left( {s + t\left( {x - s}\right) }\right) \forall t \in \left\lbrack {0,1}\right\rbrack \) ; (c) \( {d}_{S}\left( {s + t\left( {x - s}\right) }\right) = t\parallel x - s\parallel \forall t \in \left\lbrack {0,1}\right\rbrack \) ; and (d) \( \left\langle {x - s,{s}^{\prime } - s}\right\rangle \leq \frac{1}{2}{\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2}\forall {s}^{\prime } \in S \) . 1.4. Exercise. For \( 0 < t < 1 \) in Proposition 1.3(b), we have \[ {\operatorname{proj}}_{S}\left( {s + t\left( {x - s}\right) }\right) = \{ s\} \] that is, if \( x \) has a closest point \( s \) in \( S \), then \( s + t\left( {x - s}\right) \) has a unique closest point in \( S \) . (See Figure 1.2, taking \( x = {x}_{1}, s = {s}_{3} \), and \( s + t\left( {x - s}\right) = {x}_{2} \) .) The first part of the following result follows readily from the cone property of \( {N}_{S}^{P}\left( s\right) \) and the characterization (d) of Proposition 1.3; the second part demonstrates that \( P \) -normality is essentially a local property: the proximal normal cones \( {N}_{{S}_{1}}^{P}\left( s\right) \) and \( {N}_{{S}_{2}}^{P}\left( s\right) \) are the same if the two sets \( {S}_{1} \) and \( {S}_{2} \) are the same in a neighborhood of \( s \
1012_(GTM178)Nonsmooth Analysis and Control Theory
8
gle {x - s,{s}^{\prime } - s}\right\rangle \leq \frac{1}{2}{\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2}\forall {s}^{\prime } \in S \) . 1.4. Exercise. For \( 0 < t < 1 \) in Proposition 1.3(b), we have \[ {\operatorname{proj}}_{S}\left( {s + t\left( {x - s}\right) }\right) = \{ s\} \] that is, if \( x \) has a closest point \( s \) in \( S \), then \( s + t\left( {x - s}\right) \) has a unique closest point in \( S \) . (See Figure 1.2, taking \( x = {x}_{1}, s = {s}_{3} \), and \( s + t\left( {x - s}\right) = {x}_{2} \) .) The first part of the following result follows readily from the cone property of \( {N}_{S}^{P}\left( s\right) \) and the characterization (d) of Proposition 1.3; the second part demonstrates that \( P \) -normality is essentially a local property: the proximal normal cones \( {N}_{{S}_{1}}^{P}\left( s\right) \) and \( {N}_{{S}_{2}}^{P}\left( s\right) \) are the same if the two sets \( {S}_{1} \) and \( {S}_{2} \) are the same in a neighborhood of \( s \) . The inequality in Proposition 1.5(a) is called the proximal normal inequality. ## 1.5. Proposition. (a) A vector \( \zeta \) belongs to \( {N}_{S}^{P}\left( s\right) \) iff there exists \( \sigma = \sigma \left( {\zeta, s}\right) \geq 0 \) such that \[ \left\langle {\zeta ,{s}^{\prime } - s}\right\rangle \leq \sigma {\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2}\forall {s}^{\prime } \in S. \] (b) Furthermore, for any given \( \delta > 0 \), we have \( \zeta \in {N}_{S}^{P}\left( s\right) \) iff there exists \( \sigma = \sigma \left( {\zeta, s}\right) \geq 0 \) such that \[ \left\langle {\zeta ,{s}^{\prime } - s}\right\rangle \leq \sigma {\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2}\forall {s}^{\prime } \in S \cap B\left( {s;\delta }\right) . \] The only item requiring proof is the following: 1.6. Exercise. Prove that if the inequality of (b) holds for some \( \sigma \) and \( \delta \), then that of (a) holds for some possibly larger \( \sigma \) . The previous proposition makes it evident that \( {N}_{S}^{P}\left( s\right) \) is convex; however, it need be neither open nor closed. That \( {N}_{S}^{P}\left( s\right) \) can be trivial (i.e., reduce to \( \{ 0\} \) ) even when \( S \) is a closed subset of \( {\mathbb{R}}^{n} \) and \( s \) lies in bdry \( S \), can easily be seen by considering the set \[ S \mathrel{\text{:=}} \left\{ {\left( {x, y}\right) \in {\mathbb{R}}^{2} : y \geq - \left| x\right| }\right\} . \] There are no points outside \( S \) whose closest point in \( S \) is \( \left( {0,0}\right) \) (to put this another way: no ball whose interior fails to intersect \( S \) can have \( \left( {0,0}\right) \) on its boundary). Thus \( {N}_{S}^{P}\left( {0,0}\right) = \{ 0\} \) . A slightly more complicated but smoother example is the following: 1.7. Exercise. Consider \( S \) defined as \[ S \mathrel{\text{:=}} \left\{ {\left( {x, y}\right) \in {\mathbb{R}}^{2} : y \geq - {\left| x\right| }^{3/2}}\right\} . \] Show that for \( \left( {x, y}\right) \in \operatorname{bdry}S,{N}_{S}^{P}\left( {x, y}\right) = \left( {0,0}\right) \) iff \( \left( {x, y}\right) = \left( {0,0}\right) \) . 1.8. Exercise. Let \( X = {X}_{1} \oplus {X}_{2} \) be an orthogonal decomposition, and suppose \( S \subseteq X \) is closed, \( s \in S \), and \( \zeta \in {N}_{S}^{P}\left( s\right) \) . Write \( s = \) \( \left( {{s}_{1},{s}_{2}}\right) \) and \( \zeta = \left( {{\zeta }_{1},{\zeta }_{2}}\right) \) according to the given decomposition, and define \( {S}_{1} = \left\{ {{s}_{1}^{\prime } : \left( {{s}_{1}^{\prime },{s}_{2}}\right) \in S}\right\} \), and similarly define \( {S}_{2} \) . Show that \( {\zeta }_{i} \in {N}_{{S}_{i}}^{P}\left( {s}_{i}\right), i = 1,2 \) . The next two propositions illustrate that the concept of a proximal normal generalizes two classical definitions, that of a normal direction to a \( {C}^{2} \) manifold as defined in differential geometry, and that of a normal vector in the context of convex analysis. Consider a closed subset \( S \) of \( {\mathbb{R}}^{n} \) that admits a representation of the form \[ S = \left\{ {x \in {\mathbb{R}}^{n} : {h}_{i}\left( x\right) = 0, i = 1,2,\ldots, k}\right\} , \] (1) where each \( {h}_{i} : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is \( {C}^{1} \) . If the vectors \( \left\{ {\nabla {h}_{i}\left( s\right) }\right\} \left( {i = 1,2,\ldots, k}\right) \) are linearly independent at each \( s \in S \), then \( S \) is a \( {C}^{1} \) manifold of dimension \( n - k \) . 1.9. Proposition. Let \( s \in S \), where \( S \) is given by (1), and assume that the vectors \( \left\{ {\nabla {h}_{i}\left( s\right) }\right\} \left( {i = 1,2,\ldots, k}\right) \) are linearly independent. Then: (a) \( {N}_{S}^{P}\left( s\right) \subseteq \operatorname{span}\left\{ {\nabla {h}_{i}\left( s\right) }\right\} \left( {i = 1,2,\ldots, k}\right) \) . (b) If in addition each \( {h}_{i} \) is \( {C}^{2} \), then equality holds in (a). Proof. Let \( \zeta \) belong to \( {N}_{S}^{P}\left( s\right) \) . By Proposition 1.5, there exists a constant \( \sigma > 0 \) so that \[ \left\langle {\zeta ,{s}^{\prime } - s}\right\rangle \leq \sigma {\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2} \] whenever \( {s}^{\prime } \) belongs to \( S \) . Put another way, this is equivalent to saying that the point \( s \) minimizes the function \( {s}^{\prime } \mapsto \left\langle {-\zeta ,{s}^{\prime }}\right\rangle + \sigma {\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2} \) over all points \( {s}^{\prime } \) satisfying \( {h}_{i}\left( {s}^{\prime }\right) = 0\left( {i = 1,2,\ldots, k}\right) \) . The Lagrange multiplier rule of classical calculus provides a set of scalars \( {\left\{ {\mu }_{i}\right\} }_{i = 1}^{k} \) such that \( \zeta = \) \( \mathop{\sum }\limits_{i}{\mu }_{i}\nabla {h}_{i}\left( s\right) \), which establishes (a). Now let \( \zeta \) have the form \( \mathop{\sum }\limits_{i}{\mu }_{i}\nabla {h}_{i}\left( s\right) \), where each \( {h}_{i} \) is \( {C}^{2} \) . Consider the \( {C}^{2} \) function \[ g\left( x\right) \mathrel{\text{:=}} \langle - \zeta, x\rangle + \mathop{\sum }\limits_{i}{\mu }_{i}{h}_{i}\left( x\right) + \sigma \parallel x - s{\parallel }^{2}, \] where \( \sigma > 0 \) . Then \( {g}^{\prime }\left( s\right) = 0 \), and for \( \sigma \) sufficiently large we have \( {g}^{\prime \prime }\left( s\right) > 0 \) (positive definite), from which it follows that \( g \) admits a local minimum at \( s \) . Consequently, if \( {s}^{\prime } \) is near enough to \( s \) and satisfies \( {h}_{i}\left( {s}^{\prime }\right) = 0 \) for each \( i \), we have \[ g\left( {s}^{\prime }\right) = \left\langle {-\zeta ,{s}^{\prime }}\right\rangle + \sigma {\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2} \geq g\left( s\right) = \langle - \zeta, s\rangle . \] This confirms the proximal normal inequality and completes the proof. The special case in which \( S \) is convex is an important one. 1.10. Proposition. Let \( S \) be closed and convex. Then (a) \( \zeta \in {N}_{S}^{P}\left( s\right) \) iff \[ \left\langle {\zeta ,{s}^{\prime } - s}\right\rangle \leq 0\forall {s}^{\prime } \in S \] (b) If \( X \) is finite-dimensional and \( s \in \operatorname{bdry}\left( S\right) \), then \( {N}_{S}^{P}\left( s\right) \neq \{ 0\} \) . Proof. The inequality in (a) holds iff the proximal normal inequality holds with \( \sigma = 0 \) . Hence the "if" statement is immediate from Proposition 1.5(a). To see the converse, let \( \zeta \in {N}_{S}^{P}\left( s\right) \) and \( \sigma > 0 \) be chosen as in the proximal normal inequality. Let \( {s}^{\prime } \) be any point in \( S \) . Since \( S \) is convex, the point \( \widetilde{s} \mathrel{\text{:=}} s + t\left( {{s}^{\prime } - s}\right) = t{s}^{\prime } + \left( {1 - t}\right) s \) also belongs to \( S \) for each \( t \in \left( {0,1}\right) \) . The proximal normal inequality applied to \( \widetilde{s} \) gives \[ \left\langle {\zeta, t\left( {{s}^{\prime } - s}\right) }\right\rangle \leq \sigma {t}^{2}{\left| {s}^{\prime } - s\right| }^{2} \] Dividing across by \( t \) and letting \( t \downarrow 0 \) yields the desired inequality. To prove (b), let \( \left\{ {s}_{i}\right\} \) be a sequence in \( S \) converging to \( s \) so that \( {N}_{S}^{P}\left( {s}_{i}\right) \neq \) \( \{ 0\} \) for all \( i \) . Such a sequence exists by Exercise 1.2(d). Let \( {\zeta }_{i} \in {N}_{S}^{P}\left( {s}_{i}\right) \) satisfy \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} = 1 \), and passing to a subsequence if necessary, assume that \( {\zeta }_{i} \rightarrow \zeta \) as \( i \rightarrow \infty \), and note that \( \parallel \zeta \parallel = 1 \) . By part (a), we have \[ \left\langle {{\zeta }_{i},{s}^{\prime } - {s}_{i}}\right\rangle \leq 0\forall {s}^{\prime } \in S. \] Letting \( i \rightarrow \infty \) yields \[ \left\langle {\zeta ,{s}^{\prime } - s}\right\rangle \leq 0\forall {s}^{\prime } \in S \] which, again by part (a), says that \( \zeta \in {N}_{S}^{P}\left( s\right) \) . Let \( 0 \neq \zeta \in X \) and \( r \in \mathbb{R} \) . A hyperplane (with associated normal vector \( \zeta ) \) is any set of the form \( \{ x \in X : \langle \zeta, x\rangle = r\} \), and a half-space is a set of the form \( \{ x \in X : \langle \zeta, x\rangle \leq r\} \) . Proposition 1.10(b) is a separation theorem, for it says that each point in the boundary of a convex set lies in some hyperplane, with the set itself lying in one of the associated half-spaces. An example given in the end-of-chapter problems shows that this fact fails in general when \( X \) is infinite dimensional, although separation does hold under additional hypotheses. We now turn our attention from sets to functions. ## 2 Proximal Subgradients We begin by establishing some notation and recalling some facts about functions. A quite useful convention prevalent in the theories of integration and optimization, which we will also
1012_(GTM178)Nonsmooth Analysis and Control Theory
9
, again by part (a), says that \( \zeta \in {N}_{S}^{P}\left( s\right) \) . Let \( 0 \neq \zeta \in X \) and \( r \in \mathbb{R} \) . A hyperplane (with associated normal vector \( \zeta ) \) is any set of the form \( \{ x \in X : \langle \zeta, x\rangle = r\} \), and a half-space is a set of the form \( \{ x \in X : \langle \zeta, x\rangle \leq r\} \) . Proposition 1.10(b) is a separation theorem, for it says that each point in the boundary of a convex set lies in some hyperplane, with the set itself lying in one of the associated half-spaces. An example given in the end-of-chapter problems shows that this fact fails in general when \( X \) is infinite dimensional, although separation does hold under additional hypotheses. We now turn our attention from sets to functions. ## 2 Proximal Subgradients We begin by establishing some notation and recalling some facts about functions. A quite useful convention prevalent in the theories of integration and optimization, which we will also adopt, is to allow for functions \( f : X \rightarrow \) \( ( - \infty , + \infty \rbrack \) ; that is, functions which are extended real-valued. As we will see, there are many advantages in allowing \( f \) to actually attain the value \( + \infty \) at a given point. To single out those points at which \( f \) is not \( + \infty \), we define the (effective) domain as the set \[ \operatorname{dom}f \mathrel{\text{:=}} \{ x \in X : f\left( x\right) < \infty \} . \] The graph and epigraph of \( f \) are given, respectively, by \[ \operatorname{gr}f \mathrel{\text{:=}} \{ \left( {x, f\left( x\right) }\right) : x \in \operatorname{dom}f\} \] \[ \operatorname{epi}f \mathrel{\text{:=}} \{ \left( {x, r}\right) \in \operatorname{dom}f \times \mathbb{R} : r \geq f\left( x\right) \} . \] Just as sets are customarily assumed to be closed, the usual background assumption on \( f \) is that of lower semicontinuity. A function \( f : X \rightarrow \) \( ( - \infty , + \infty \rbrack \) is lower semicontinuous at \( x \) provided that \[ \mathop{\liminf }\limits_{{{x}^{\prime } \rightarrow x}}f\left( {x}^{\prime }\right) \geq f\left( x\right) \] This condition is clearly equivalent to saying that for all \( \varepsilon > 0 \), there exists \( \delta > 0 \) so that \( y \in B\left( {x;\delta }\right) \) implies \( f\left( y\right) \geq f\left( x\right) - \varepsilon \), where as usual, \( \infty - r \) is interpreted as \( \infty \) when \( r \in \mathbb{R} \) . Complementary to lower semicontinuity is upper semicontinuity: \( f \) is upper semicontinuous at \( x \) if \( - f \) is lower semicontinuous at \( x \) . Lower semicontinuous functions are featured prominently in our development, but of course our results have upper semicontinuous analogues, although we will rarely state them. This preference for lower semicontinuity explains why \( + \infty \) is allowed as a function value and not \( - \infty \) . As is customary, we say that a function \( f \) is continuous at \( x \in X \) provided it is finite-valued near \( x \) and for all \( \varepsilon > 0 \), there exists \( \delta > 0 \) so that \( y \in B\left( {x;\delta }\right) \) implies \( \left| {f\left( x\right) - f\left( y\right) }\right| \leq \varepsilon \) . For finite-valued \( f \), this is equivalent to saying that \( f \) is both lower and upper semicontinuous at \( x \) . If \( f \) is lower semicontinuous (respectively, upper semicontinuous, continuous) at each point \( x \) in an open set \( U \subset X \), then \( f \) is called lower semicontinuous (respectively, upper semicontinuous, continuous) on \( U \) . To restrict certain pathological functions from entering the discussion, we designate by \( \mathcal{F}\left( U\right) \), where \( U \subseteq X \) is open, the class of all functions \( f : X \rightarrow \) \( ( - \infty ,\infty \rbrack \) which are lower semicontinuous on \( U \) and such that \( \operatorname{dom}f \cap U \neq \varnothing \) . If \( U = X \), then we simply write \( \mathcal{F} \) for \( \mathcal{F}\left( X\right) \) . Let \( S \) be a subset of \( X \) . The indicator function of \( S \), denoted either by \( {I}_{S}\left( \cdot \right) \) or \( I\left( {\cdot ;S}\right) \), is the extended-valued function defined by \[ {I}_{S}\left( x\right) \mathrel{\text{:=}} \left\{ \begin{array}{ll} 0 & \text{ if }x \in S \\ + \infty & \text{ otherwise. } \end{array}\right. \] Let \( U \subset X \) be an open convex set. A function \( f : X \rightarrow ( - \infty ,\infty \rbrack \) is said to be convex on \( U \) provided \[ f\left( {{tx} + \left( {1 - t}\right) y}\right) \leq {tf}\left( x\right) + \left( {1 - t}\right) f\left( y\right) \forall x, y \in U,0 < t < 1. \] A function \( f \) which is convex on \( X \) is simply said to be convex. Note that dom \( f \) is necessarily a convex set if \( f \) is convex. The following exercise contains some elementary properties of lower semicontinuous and convex functions. Parts (a) and (b) in particular help to demonstrate why the epigraph, rather than the graph, of a function plays the fundamental role in the analysis of lower semicontinuous functions. Note that \( X \times \mathbb{R} \), the space in which epi \( f \) lives, is always viewed as a Hilbert space with inner product \( \left\langle {\left( {x, r}\right) ,\left( {{x}^{\prime },{r}^{\prime }}\right) }\right\rangle \mathrel{\text{:=}} \left\langle {x,{x}^{\prime }}\right\rangle + r{r}^{\prime } \) . 2.1. Exercise. Suppose \( f : X \rightarrow ( - \infty , + \infty \rbrack \) . (a) Show that \( f \) is lower semicontinuous on \( X \) iff epi \( f \) is closed in \( X \times \mathbb{R} \), and this is true iff each \( r \) -level set \( \{ x : f\left( x\right) \leq r\} \) is closed, \( r \in \mathbb{R} \) . Note that \( \operatorname{gr}f \) need not be closed when \( f \) is lower semicontinuous. (b) Show that \( f \) is convex on \( X \) iff epi \( f \) is a convex subset of \( X \times \mathbb{R} \) . (c) When \( f \) is an indicator function, \( f = {I}_{S} \), then \( f \in \mathcal{F}\left( X\right) \) iff \( S \) is nonempty and closed, and \( f \) is convex iff \( S \) is convex. (d) Suppose that \( \left( {\zeta , - \lambda }\right) \in X \times \mathbb{R} \) belongs to \( {N}_{\text{epi }f}^{P}\left( {x, r}\right) \) for some \( \left( {x, r}\right) \in \operatorname{epi}f \), where \( f \in \mathcal{F} \) . Prove that \( \lambda \geq 0 \), that \( r = f\left( x\right) \) if \( \lambda > 0 \), and that \( \lambda = 0 \) if \( r > f\left( x\right) \) . In this last case, show that \( \left( {\zeta ,0}\right) \in {N}_{\text{epi }f}^{P}\left( {x, f\left( x\right) }\right) . \) (e) Give an example of a continuous \( f \in \mathcal{F}\left( \mathbb{R}\right) \) such that at some point \( x \) we have \( \left( {1,0}\right) \in {N}_{\text{epi }f}^{P}\left( {x, f\left( x\right) }\right) \) . (f) If \( S = \operatorname{epi}f \), where \( f \in \mathcal{F} \), prove that for all \( x,{d}_{S}\left( {x, r}\right) \) is nonincreasing as a function of \( r \) . A vector \( \zeta \in X \) is called a proximal subgradient (or \( P \) -subgradient) of a lower semicontinuous function \( f \) at \( x \in \operatorname{dom}f \) provided that \[ \left( {\zeta , - 1}\right) \in {N}_{\text{epi }f}^{P}\left( {x, f\left( x\right) }\right) . \] The set of all such \( \zeta \) is denoted \( {\partial }_{P}f\left( x\right) \), and is referred to as the proximal subdifferential, or \( P \) -subdifferential. Note that because a cone is involved, if \( \alpha > 0 \) and \( \left( {\zeta , - \alpha }\right) \in {N}_{\text{epi }f}^{P}\left( {x, f\left( x\right) }\right) \), then \( \zeta /\alpha \in {\partial }_{P}f\left( x\right) \) . It also follows immediately from our study of the proximal normal cone that \( {\partial }_{P}f\left( x\right) \) is convex, however it is not necessarily open, closed, or nonempty. The function \( f : \mathbb{R} \rightarrow \mathbb{R} \) defined by \( f\left( x\right) = - \left| x\right| \) is a simple example of a continuous function having \( {\partial }_{P}f\left( 0\right) = \varnothing \) . Figure 1.3 illustrates the epigraph of a function \( f \) together with some vectors of the form \( \left( {\zeta , - 1}\right) ,\zeta \in {\partial }_{P}f\left( x\right) \) . There exists a single \( P \) -subgradient at \( {x}_{1} \), as well as at all the unlabeled points. At \( {x}_{2} \), there are no \( P \) - subgradients, and there are multiple \( P \) -subgradients at the three remaining labeled points. At \( {x}_{4} \), the proximal subdifferential is an unbounded set; the (horizontal) dashed arrow here is not associated with a \( P \) -subgradient, although it does represent a \( P \) -normal to epi \( f \) . The indicator function is one of several ways in which we pass between sets and functions. It is also useful in optimization: note that minimizing \( f \) over a set \( S \) is equivalent to minimizing the function \( f + {I}_{S} \) globally. ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_42_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_42_0.jpg) FIGURE 1.3. The epigraph of a function. 2.2. Exercise. Let \( f = {I}_{S} \) . Prove that for \( x \in S \) we have \[ {\partial }_{P}f\left( x\right) = {\partial }_{P}{I}_{S}\left( x\right) = {N}_{S}^{P}\left( x\right) . \] The main theme of this chapter is to develop the calculus rules governing the proximal subgradient. We will see that to a surprising degree, many of the usual properties enjoyed by the classical derivative carry over to the proximal subgradient \( {\partial }_{P}f\left( x\right) \) . As a first illustration of this, we give an exercise which echos the vanishing of the derivative at a local minimum. A point \( x \in X \) is said to attain the minimum of \( f \) on \( S \) provided \( x \in \) \( S \cap \operatorname{dom}f \) and \[ f\left( x\right) \leq f\left( y\right) \forall y \in S. \] If there exists an open neighborhood \( U \) of \
1012_(GTM178)Nonsmooth Analysis and Control Theory
10
I}_{S} \) globally. ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_42_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_42_0.jpg) FIGURE 1.3. The epigraph of a function. 2.2. Exercise. Let \( f = {I}_{S} \) . Prove that for \( x \in S \) we have \[ {\partial }_{P}f\left( x\right) = {\partial }_{P}{I}_{S}\left( x\right) = {N}_{S}^{P}\left( x\right) . \] The main theme of this chapter is to develop the calculus rules governing the proximal subgradient. We will see that to a surprising degree, many of the usual properties enjoyed by the classical derivative carry over to the proximal subgradient \( {\partial }_{P}f\left( x\right) \) . As a first illustration of this, we give an exercise which echos the vanishing of the derivative at a local minimum. A point \( x \in X \) is said to attain the minimum of \( f \) on \( S \) provided \( x \in \) \( S \cap \operatorname{dom}f \) and \[ f\left( x\right) \leq f\left( y\right) \forall y \in S. \] If there exists an open neighborhood \( U \) of \( x \in X \) on which \( x \) attains the minimum of \( f \), then \( x \) is said to be a local minimum of \( f \) . If \( x \) is a minimum of \( f \) on \( U = X \), then \( x \) is called a global minimum. 2.3. Exercise. Suppose \( f \in \mathcal{F} \) . (a) Show that if \( f \) attains a local minimum at \( x \), then \( 0 \in {\partial }_{P}f\left( x\right) \) . (b) Suppose \( S \subset X \) is compact and satisfies \( S \cap \operatorname{dom}f \neq \varnothing \) . Show that \( f \) is bounded below on \( S \), and attains its minimum over \( S \) . ## Classical Derivatives Before developing further properties of \( P \) -subgradients, we need to recall some facts about classical derivatives. We will do so rather quickly. The directional derivative of \( f \) at \( x \in \operatorname{dom}f \) in the direction \( v \in X \) is defined as \[ {f}^{\prime }\left( {x;v}\right) \mathrel{\text{:=}} \mathop{\lim }\limits_{{t \downarrow 0}}\frac{f\left( {x + {tv}}\right) - f\left( x\right) }{t}, \] (1) when the limit exists. We say that \( f \) is Gâteaux differentiable at \( x \) provided the limit in (2.1) exists for all \( v \in X \), and there exists a (necessarily unique) element \( {f}_{G}^{\prime }\left( x\right) \in X \) (called the Gâteaux derivative) that satisfies \[ {f}^{\prime }\left( {x;v}\right) = \left\langle {{f}_{G}^{\prime }\left( x\right), v}\right\rangle \forall v \in X. \] (2) A function may possess a directional derivative at \( x \) in every direction and yet fail to possess a Gâteaux derivative, as is evidenced by \( f\left( x\right) = \parallel x\parallel \) at \( x = 0 \) . In this case, we have \( {f}^{\prime }\left( {0;v}\right) = \parallel v\parallel \) . Also, a lower semicontinuous function may have a Gâteaux derivative at a point \( x \) but not be continuous there. Suppose that (2) holds at a point \( x \), and in addition that the convergence in (1) is uniform with respect to \( v \) in bounded subsets of \( X \) . We then say that \( f \) is Fréchet differentiable at \( x \), and in this case write \( {f}^{\prime }\left( x\right) \) (the Fréchet derivative) in place of \( {f}_{G}^{\prime }\left( x\right) \) . Equivalently this means that for all \( r > 0 \) and \( \varepsilon > 0 \), there exists \( \delta > 0 \) so that \[ \left| {\frac{f\left( {x + {tv}}\right) - f\left( x\right) }{t} - \left\langle {{f}^{\prime }\left( x\right), v}\right\rangle }\right| < \varepsilon \] holds for all \( \left| t\right| < \delta \) and \( \parallel v\parallel \leq r \) . The two notions of differentiability are not equivalent, even in finite dimensions. We can easily show that Fréchet differentiability at \( x \) implies continuity at \( x \), which is not the case for Gâteaux differentiability. Many of the elementary properties of the derivative encountered in the multivariate calculus (i.e., when \( X = {\mathbb{R}}^{n} \) ) have exact analogues using either Fréchet or Gâteaux derivatives, where \( {f}^{\prime } \) or \( {f}_{G}^{\prime } \) take the place of the usual gradient \( \nabla f \) . To illustrate in some detail, suppose \( f, g : X \rightarrow \mathbb{R} \) have Fréchet derivatives at \( x \in X \) . Then \( f \pm g,{fg} \), and \( f/g \) (with \( g\left( x\right) \neq 0 \) ) all have Fréchet derivatives at \( x \) obeying the classical rules: \[ {\left( f \pm g\right) }^{\prime }\left( x\right) = {f}^{\prime }\left( x\right) \pm {g}^{\prime }\left( x\right) \] \[ {\left( fg\right) }^{\prime }\left( x\right) = {f}^{\prime }\left( x\right) g\left( x\right) + f\left( x\right) {g}^{\prime }\left( x\right) , \] \[ {\left( \frac{f}{g}\right) }^{\prime }\left( x\right) = \left( \frac{{f}^{\prime }\left( x\right) g\left( x\right) - f\left( x\right) {g}^{\prime }\left( x\right) }{{g}^{2}\left( x\right) }\right) . \] The proofs of these facts are the same as in the classical case. The Mean Value Theorem can be stated as follows: suppose \( f \in \mathcal{F}\left( X\right) \) is Gâteaux differentiable on an open neighborhood that contains the line segment \( \left\lbrack {x, y}\right\rbrack \mathrel{\text{:=}} \{ {tx} + \left( {1 - t}\right) y : 0 \leq t \leq 1\} \), where \( x, y \in X \) . That is, there exists an open set \( U \) containing the line segment \( \left\lbrack {x, y}\right\rbrack \) such that \( f \) is differentiable at every point of \( U \) . Then there exists a point \( z \mathrel{\text{:=}} {tx} + \left( {1 - t}\right) y \) , \( 0 < t < 1 \), so that \[ f\left( y\right) - f\left( x\right) = \left\langle {{f}_{G}^{\prime }\left( z\right), y - x}\right\rangle . \] A proof of the Mean Value Theorem can be given by applying the classical one-dimensional mean value theorem to the function \( g : \left\lbrack {0,1}\right\rbrack \rightarrow \mathbb{R} \) defined by \( g\left( t\right) = f\left( {x + t\left( {y - x}\right) }\right) \) . Another useful result is the Chain Rule. In order to state it, we first need to extend the above notions of differentiability to maps between two Hilbert spaces. Suppose \( {X}_{1} \) and \( {X}_{2} \) are Hilbert spaces with norms \( \parallel \cdot {\parallel }_{1} \) and \( \parallel \cdot {\parallel }_{2} \) , respectively, and suppose \( F : {X}_{1} \rightarrow {X}_{2} \) is a mapping between these spaces. We write \( \mathcal{L}\left( {{X}_{1},{X}_{2}}\right) \) for the space of bounded linear transformations from \( {X}_{1} \) to \( {X}_{2} \) endowed with the usual operator norm. The scalar case \( {X}_{2} = \mathbb{R} \) was discussed above, in which case \( \mathcal{L}\left( {{X}_{1},\mathbb{R}}\right) \) was identified with \( {X}_{1} \) in the usual way. Let \( x \in {X}_{1} \) . The Gâteaux derivative, should it exist, of \( F \) at \( x \) is an element \( {F}_{G}^{\prime }\left( x\right) \in \mathcal{L}\left( {{X}_{1},{X}_{2}}\right) \) that satisfies \[ \mathop{\lim }\limits_{{t \downarrow 0}}{\begin{Vmatrix}\frac{F\left( {x + {tv}}\right) - F\left( x\right) }{t} - {F}_{G}^{\prime }\left( x\right) \left( v\right) \end{Vmatrix}}_{2} = 0, \] for all \( v \in {X}_{1} \) . Should in addition the above limit hold uniformly over \( v \) in bounded sets of \( {X}_{1} \), then \( F \) is Fréchet differentiable and we write \( {F}^{\prime }\left( x\right) \) in place of \( {F}_{G}^{\prime }\left( x\right) \) . As in the scalar case, the derivative of the sum of two functions mapping \( {X}_{1} \) to \( {X}_{2} \) is the sum of the derivatives. Let us now consider the Chain Rule. Suppose \( {X}_{1},{X}_{2} \), and \( {X}_{3} \) are all Hilbert spaces, and \( F : {X}_{1} \rightarrow {X}_{2} \) , \( G : {X}_{2} \rightarrow {X}_{3} \) . Assume that \( F \) is Fréchet differentiable at \( x \in {X}_{1} \), and \( G \) is Fréchet differentiable at \( F\left( x\right) \in {X}_{2} \) . Then the composition \( G \circ F : {X}_{1} \rightarrow {X}_{3} \) is Fréchet differentiable at \( x \) and \[ {\left( G \circ F\right) }^{\prime }\left( x\right) = {G}^{\prime }\left( {F\left( x\right) }\right) {F}^{\prime }\left( x\right) \] where \( {G}^{\prime }\left( {F\left( x\right) }\right) {F}^{\prime }\left( x\right) \in \mathcal{L}\left( {{X}_{1},{X}_{3}}\right) \) signifies the composition of \( {F}^{\prime }\left( x\right) \) with \( {G}^{\prime }\left( {F\left( x\right) }\right) \) . Suppose \( U \subseteq X \) is open and \( f : U \rightarrow \mathbb{R} \) is Fréchet differentiable on \( U \) . If \( {f}^{\prime }\left( \cdot \right) : U \rightarrow X \) is continuous on \( U \), then we say that \( f \) is \( {C}^{1} \) on \( U \), and write \( f \in {C}^{1}\left( U\right) \) . It turns out that if \( f \) is Gâteaux differentiable on \( U \) with a continuous derivative there, then \( f \in {C}^{1}\left( U\right) \) . Now suppose further that the map \( {f}^{\prime }\left( \cdot \right) : U \rightarrow X \) is itself Fréchet differentiable on \( U \) with its derivative at \( x \in U \) denoted by \( {f}^{\prime \prime }\left( x\right) \in \mathcal{L}\left( {X, X}\right) \) (in the multivariate calculus, \( {f}^{\prime \prime }\left( x\right) \) is the Hessian). For each \( x \in U, f \) then admits a local second-order Taylor expansion with remainder, which means there exists a neighborhood \( B\left( {x;\eta }\right) \) of \( x \) so that for every \( y \in B\left( {x;\eta }\right) \) we have \[ f\left( y\right) = f\left( x\right) + \left\langle {{f}^{\prime }\left( x\right), y - x}\right\rangle + \frac{1}{2}\left\langle {{f}^{\prime \prime }\left( z\right) \left( {y - x}\right), y - x}\right\rangle , \] where \( z \) is some element on the line segment connecting \( x \) and \( y \) . We note that if the norms of \( {f}^{\prime \prime }\left( y\right) \) are bounded over \( y \in B\left( {x;\eta }\right) \) by the constant \( {2\sigma } > 0 \), then this implies \[ f\left( y\right) \geq f\left( x\right) + \left\langle {{f}^{\prime }\left( x\right), y - x}\right\rangle - \sigma \parallel y - x{\parallel }^{2} \] (3) for a
1012_(GTM178)Nonsmooth Analysis and Control Theory
11
L}\left( {X, X}\right) \) (in the multivariate calculus, \( {f}^{\prime \prime }\left( x\right) \) is the Hessian). For each \( x \in U, f \) then admits a local second-order Taylor expansion with remainder, which means there exists a neighborhood \( B\left( {x;\eta }\right) \) of \( x \) so that for every \( y \in B\left( {x;\eta }\right) \) we have \[ f\left( y\right) = f\left( x\right) + \left\langle {{f}^{\prime }\left( x\right), y - x}\right\rangle + \frac{1}{2}\left\langle {{f}^{\prime \prime }\left( z\right) \left( {y - x}\right), y - x}\right\rangle , \] where \( z \) is some element on the line segment connecting \( x \) and \( y \) . We note that if the norms of \( {f}^{\prime \prime }\left( y\right) \) are bounded over \( y \in B\left( {x;\eta }\right) \) by the constant \( {2\sigma } > 0 \), then this implies \[ f\left( y\right) \geq f\left( x\right) + \left\langle {{f}^{\prime }\left( x\right), y - x}\right\rangle - \sigma \parallel y - x{\parallel }^{2} \] (3) for all \( y \in B\left( {x;\eta }\right) \) . If it should also happen that \( {f}^{\prime \prime } : X \rightarrow \mathcal{L}\left( {X, X}\right) \) is continuous on \( U \), then \( f \) is said to be twice continuously differentiable on \( U \), and we write \( f \in {C}^{2}\left( U\right) \) , or simply \( f \in {C}^{2} \) if \( U = X \) . We note that if \( f \in {C}^{2}\left( U\right) \), then for each \( x \in U \) there exists a neighborhood \( B\left( {x;\eta }\right) \) and a constant \( \sigma \) so that (3) holds, since the continuity of \( {f}^{\prime \prime } \) at \( x \) implies that the norms of \( {f}^{\prime \prime } \) are bounded in a neighborhood of \( x \) . ## 2.4. Exercise. (a) Let \( x \in X \) and define \( f : X \rightarrow \mathbb{R} \) by \( f\left( y\right) = \parallel y - x{\parallel }^{2} \) . Show that \( f \in {C}^{2} \), and that for each \( y \in X \), we have \( {f}^{\prime }\left( y\right) = 2\left( {y - x}\right) \) and \( {f}^{\prime \prime }\left( y\right) = 2\mathcal{I} \), where \( \mathcal{I} \in \mathcal{L}\left( {X, X}\right) \) is the identity transformation. (b) Suppose \( c > 0 \) is a constant, and \( x \) and \( \zeta \) are fixed elements in \( X \) . Define \( g : X \rightarrow \mathbb{R} \) by \[ g\left( y\right) = {\left\lbrack {c}^{2} + 2c\langle \zeta, y - x\rangle - \parallel y - x{\parallel }^{2}\right\rbrack }^{1/2}. \] Show that \( g \in {C}^{2}\left( U\right) \) for some neighborhood \( U \) of \( x \), and that \( {g}^{\prime }\left( x\right) = \zeta \) . (c) Let \( f\left( x\right) = \parallel x\parallel \) . Then \( {f}^{\prime }\left( x\right) \) exists for \( x \neq 0 \), and equals \( x/\parallel x\parallel \) . We now return to developing properties of the proximal subgradient. The following characterization is the most widely used description of the proximal subgradient, and we give it a name: by the proximal subgradient inequality, we mean the inequality appearing in the following result: 2.5. Theorem. Let \( f \in \mathcal{F} \) and let \( x \in \operatorname{dom}\left( f\right) \) . Then \( \zeta \in {\partial }_{P}f\left( x\right) \) if and only if there exist positive numbers \( \sigma \) and \( \eta \) such that \[ f\left( y\right) \geq f\left( x\right) + \langle \zeta, y - x\rangle - \sigma \parallel y - x{\parallel }^{2}\forall y \in B\left( {x;\eta }\right) . \] (4) Proof. Let us first prove the "if" part of the theorem's statement. The inequality (4) implies that \[ \alpha - f\left( x\right) + \sigma \left\lbrack {\parallel y - x{\parallel }^{2} + {\left( \alpha - f\left( x\right) \right) }^{2}}\right\rbrack \geq \langle \zeta, y - x\rangle \] for all \( y \in B\left( {x;\eta }\right) \) and for all \( \alpha \geq f\left( y\right) \) . This in turn implies \[ \langle \left( {\zeta , - 1}\right) ,\left\lbrack {\left( {y,\alpha }\right) - \left( {x, f\left( x\right) }\right) }\right\rbrack \rangle \leq \sigma \parallel \left( {y,\alpha }\right) - \left( {x, f\left( x\right) }\right) {\parallel }^{2} \] for all points \( \left( {y,\alpha }\right) \in \operatorname{epi}\left( f\right) \) near \( \left( {x, f\left( x\right) }\right) \) . In view of Proposition 1.5, this implies that \( \left( {\zeta , - 1}\right) \in {N}_{\text{epi }f}^{P}\left( {x, f\left( x\right) }\right) \) . Let us now turn to the "only if" part. To this end, suppose that \( \left( {\zeta , - 1}\right) \in \) \( {N}_{\text{epi }f}^{P}\left( {x, f\left( x\right) }\right) \) . Then by Proposition 1.3 there exists \( \delta > 0 \) such that \[ \left( {x, f\left( x\right) }\right) \in {\operatorname{proj}}_{\text{epi }f}\left( {\left( {x, f\left( x\right) }\right) + \delta \left( {\zeta , - 1}\right) }\right) . \] This evidently implies \[ \parallel \delta \left( {\zeta , - 1}\right) {\parallel }^{2} \leq \parallel \left\lbrack {\left( {x, f\left( x\right) }\right) + \delta \left( {\zeta , - 1}\right) }\right\rbrack - \left( {y,\alpha }\right) {\parallel }^{2} \] for all \( \left( {y,\alpha }\right) \in \operatorname{epi}f \) ; see Figure 1.4. Upon taking \( \alpha = f\left( y\right) \), the last inequality yields \[ {\delta }^{2}\parallel \zeta {\parallel }^{2} + {\delta }^{2} \leq \parallel x - y + {\delta \zeta }{\parallel }^{2} + {\left( f\left( x\right) - f\left( y\right) - \delta \right) }^{2}, \] which can be rewritten as \[ {\left( f\left( y\right) - f\left( x\right) + \delta \right) }^{2} \geq {\delta }^{2} + {2\delta }\langle \zeta, y - x\rangle - \parallel x - y{\parallel }^{2}. \] (5) It is clear that the right-hand side of (5) is positive for all \( y \) sufficiently near \( x \), say for \( y \in B\left( {x;\eta }\right) \) . By shrinking \( \eta > 0 \) if necessary, we can also ensure (by the lower semicontinuity of \( f \) ) that \( y \in B\left( {x;\eta }\right) \) implies \[ f\left( y\right) - f\left( x\right) + \delta > 0. \] Hence taking square roots of (5) gives us that \[ f\left( y\right) \geq g\left( y\right) \mathrel{\text{:=}} f\left( x\right) - \delta + {\left\{ {\delta }^{2} + 2\delta \langle \zeta, y - x\rangle - \parallel x - y{\parallel }^{2}\right\} }^{1/2} \] (6) for all \( y \in B\left( {x;\eta }\right) \) . Direct calculations show that \( {g}^{\prime }\left( x\right) = \zeta \) and that \( {g}^{\prime \prime } \) exists and is bounded, say by \( {2\sigma } > 0 \), on a neighborhood of \( x \) (Exercise 2.4). Again if \( \eta \) is shrunk further if necessary, we have (as noted above in connection with the inequality (3)) \[ g\left( y\right) \geq g\left( x\right) + \langle \zeta, y - x\rangle - \sigma \parallel y - x{\parallel }^{2}\forall y \in B\left( {x;\eta }\right) . \] But then by (6), and since \( f\left( x\right) = g\left( x\right) \), we see that \[ f\left( y\right) \geq f\left( x\right) + \langle \zeta, y - x\rangle - \sigma \parallel y - x{\parallel }^{2}\forall y \in B\left( {x;\eta }\right) , \] which is (4) as required. ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_47_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_47_0.jpg) FIGURE 1.4. \( \zeta \) belongs to \( {\partial }_{P}f\left( x\right) \) . The definition of proximal subgradients via proximal normals to an epigraph is a geometric approach, and the characterization in Theorem 2.5 can also be interpreted geometrically. The proximal subgradient inequality (4) asserts that near \( x, f\left( \cdot \right) \) majorizes the quadratic function \[ h\left( y\right) \mathrel{\text{:=}} f\left( x\right) + \langle \zeta, y - x\rangle - \sigma \parallel y - x{\parallel }^{2}, \] with equality at \( y = x \) (since obviously \( h\left( x\right) = f\left( x\right) \) ). It is worth noting that this is equivalent to saying that \( y \mapsto f\left( y\right) - h\left( y\right) \) has a local minimum at \( y = x \) with min value equal to 0 . Put into purely heuristic terms, the content of Theorem 2.5 is that the existence of such a parabola \( h \) which "locally fits under" the epigraph of \( f \) at \( \left( {x, f\left( x\right) }\right) \) is equivalent to the existence of a ball in \( X \times \mathbb{R} \) touching the epigraph nonhorizontally at that point; this is, in essence, what the proof of the theorem shows. See Figure 1.4. The description of proximal subgradients contained in Theorem 2.5 is generally more useful in analyzing lower semicontinuous functions than is a direct appeal to the definition. The first corollary below illustrates this, and relates \( {\partial }_{P}f \) to classical differentiability. It also states that for convex functions, the inequality (4) holds globally in an even simpler form; this is the functional analogue of the simplified proximal normal inequality for convex sets (Proposition 1.10). 2.6. Corollary. Let \( f \in \mathcal{F} \) and \( U \subset X \) be open. (a) Assume that \( f \) is Gâteaux differentiable at \( x \in U \) . Then \[ {\partial }_{P}f\left( x\right) \subseteq \left\{ {{f}_{G}^{\prime }\left( x\right) }\right\} \] (b) If \( f \in {C}^{2}\left( U\right) \), then \[ {\partial }_{P}f\left( x\right) = \left\{ {{f}^{\prime }\left( x\right) }\right\} \] for all \( x \in U \) . (c) If \( f \) is convex, then \( \zeta \in {\partial }_{P}f\left( x\right) \) iff \[ f\left( y\right) \geq f\left( x\right) + \langle \zeta, y - x\rangle \forall y \in X. \] (7) Proof. (a) Suppose \( f \) has a Gâteaux derivative at \( x \) and that \( \zeta \in {\partial }_{P}f\left( x\right) \) . For any \( v \in X \), if we write \( y = x + {tv} \), the proximal subgradient inequality (4) implies that there exists \( \sigma > 0 \) such that \[ \frac{f\left( {x + {tv}}\right) - f\left( x\right) }{t} - \langle \zeta, v\rangle \geq - {t\sigma }\parallel v{\parallel }^{2} \] for all sufficiently small positive \( t \) . Upon letting \( t \downarrow 0 \) we obtain \[ \left\langle {{f}_{G}^{\prime }\left( x\right) - \zeta, v}\right\rangle \geq 0 \] Since \( v \) was arbitrary, the conclusion \( \zeta = {f}_{G}^{\prime }\lef
1012_(GTM178)Nonsmooth Analysis and Control Theory
12
If \( f \in {C}^{2}\left( U\right) \), then \[ {\partial }_{P}f\left( x\right) = \left\{ {{f}^{\prime }\left( x\right) }\right\} \] for all \( x \in U \) . (c) If \( f \) is convex, then \( \zeta \in {\partial }_{P}f\left( x\right) \) iff \[ f\left( y\right) \geq f\left( x\right) + \langle \zeta, y - x\rangle \forall y \in X. \] (7) Proof. (a) Suppose \( f \) has a Gâteaux derivative at \( x \) and that \( \zeta \in {\partial }_{P}f\left( x\right) \) . For any \( v \in X \), if we write \( y = x + {tv} \), the proximal subgradient inequality (4) implies that there exists \( \sigma > 0 \) such that \[ \frac{f\left( {x + {tv}}\right) - f\left( x\right) }{t} - \langle \zeta, v\rangle \geq - {t\sigma }\parallel v{\parallel }^{2} \] for all sufficiently small positive \( t \) . Upon letting \( t \downarrow 0 \) we obtain \[ \left\langle {{f}_{G}^{\prime }\left( x\right) - \zeta, v}\right\rangle \geq 0 \] Since \( v \) was arbitrary, the conclusion \( \zeta = {f}_{G}^{\prime }\left( x\right) \) follows. (b) If \( f \in {C}^{2}\left( U\right) \) and \( x \in U \), then we have \( {f}^{\prime }\left( x\right) \in {\partial }_{P}f\left( x\right) \) by Theorem 2.5, since (3) implies (4) if \( \zeta \) is set equal to \( {f}^{\prime }\left( x\right) \) . That \( {\partial }_{P}f\left( x\right) \) contains only \( {f}^{\prime }\left( x\right) \) follows from part (a). (c) Obviously if \( \zeta \) satisfies (7), then (4) holds with \( \sigma = 0 \) and any \( \eta > 0 \) , so that \( \zeta \in {\partial }_{P}f\left( x\right) \) . Conversely, suppose \( \zeta \in {\partial }_{P}f\left( x\right) \), and \( \sigma \) and \( \eta \) are chosen as in (4). Let \( y \in X \) . Then for any \( t \) in \( \left( {0,1}\right) \) sufficiently small so that \( \left( {1 - t}\right) x + {ty} \in B\left( {x;\eta }\right) \), we have by the convexity of \( f \) and (4) (where we substitute \( \left( {1 - t}\right) x + {ty} \) for \( y \) ) that \[ \left( {1 - t}\right) f\left( x\right) + {tf}\left( y\right) \geq f\left( {\left( {1 - t}\right) x + {ty}}\right) \] \[ \geq f\left( x\right) + t\langle \zeta, y - x\rangle - {t}^{2}\sigma \parallel y - x{\parallel }^{2}. \] Simplifying and dividing by \( t \), we conclude \[ f\left( y\right) \geq f\left( x\right) + \langle \zeta, y - x\rangle - {t\sigma }\parallel y - x{\parallel }^{2}. \] Letting \( t \downarrow 0 \) yields (7). The containment in Corollary 2.6(a) is the best possible conclusion under the stated assumptions, since even when \( X = \mathbb{R} \) and \( f \) is continuously differentiable, the nonemptiness of the proximal subdifferential is not assured. The already familiar \( {C}^{1} \) function \( f\left( x\right) = - {\left| x\right| }^{3/2} \) admits no proximal subgradient at \( x = 0 \) (see Exercise 1.7). The first part of the following corollary has already been observed (Exercise 2.3). Despite its simplicity, it is the fundamental fact that generates proximal subgradients on many occasions. The second part says that the "first-order" necessary condition for a minimum is also sufficient in the case of convex functions, which is a principal reason for their importance. 2.7. Corollary. Suppose \( f \in \mathcal{F} \) . (a) If \( f \) has a local minimum at \( x \), then \( 0 \in {\partial }_{P}f\left( x\right) \) . (b) Conversely, if \( f \) is convex and \( 0 \in {\partial }_{P}f\left( x\right) \), then \( x \) is a global minimum of \( f \) . Proof. (a) The definition of a local minimum says there exists \( \eta > 0 \) so that \[ f\left( y\right) \geq f\left( x\right) \forall y \in B\left( {x;\eta }\right) \] which is the proximal subgradient inequality with \( \zeta = 0 \) and \( \sigma = 0 \) . Thus Theorem 2.5 implies that \( 0 \in {\partial }_{P}f\left( x\right) \) . (b) Under these hypotheses,(7) holds with \( \zeta = 0 \) . Thus \( f\left( y\right) \geq f\left( x\right) \) for all \( y \in X \), which says that \( x \) is a global minimum of \( f \) . The proximal subdifferential is a "one-sided" object suitable to the analysis of lower semicontinuous functions. For a theory applicable to upper semicontinuous functions \( f \), the proximal superdifferential \( {\partial }^{P}f\left( x\right) \) is the appropriate object, and can be defined simply as \( - {\partial }_{P}\left( {-f}\right) \left( x\right) \) . In the subsequent development, analogues for upper semicontinuous functions will usually not be stated because they require only evident modifications, such as replacing "sub" by "super," "≤ " by "≥," "minimum" by "maximum," and "convex" by "concave." Nonetheless, we will have occasional use for supergradients. ## 2.8. Exercise. (a) Suppose \( - f \in \mathcal{F} \) and \( x \in \operatorname{dom}\left( {-f}\right) \) . Show that \( \zeta \in {\partial }^{P}f\left( x\right) \) iff there exist positive numbers \( \sigma \) and \( \eta \) so that \[ f\left( y\right) - \langle \zeta, y - x\rangle - \sigma \parallel y - x{\parallel }^{2} \leq f\left( x\right) \forall y \in B\left( {x;\eta }\right) . \] (b) Suppose \( U \subset X \) is open, \( x \in U, f : U \rightarrow \mathbb{R} \) is continuous on \( U \) , and that both \( {\partial }_{P}f\left( x\right) \) and \( {\partial }^{P}f\left( x\right) \) are nonempty. Prove that \( f \) is Fréchet differentiable at \( x \), and that we have \( {\partial }_{P}f\left( x\right) = \) \( \left\{ {{f}^{\prime }\left( x\right) }\right\} = {\partial }^{P}f\left( x\right) . \) (c) Suppose \( f \in \mathcal{F} \) is convex and continuous at \( x \in \operatorname{int}\operatorname{dom}f \) . Show that \( {\partial }_{P}f\left( x\right) \neq \varnothing \) . (Hint. Apply the Separation Theorem (see, e.g., Rudin (1973)) to \( \left( {x, f\left( x\right) }\right) \in \) bdry epi \( f \) ). Deduce further that if \( {\partial }^{P}f\left( x\right) \neq \varnothing \), then \( f \) is Fréchet differentiable at \( x \) . There is a natural way to define partial proximal subgradients. Suppose \( X \) is decomposed into orthogonal subspaces \( X = {X}_{1} \oplus {X}_{2} \), and \( f \in \mathcal{F} \) . Let \( x \in X \), and write \( x = \left( {{x}_{1},{x}_{2}}\right) \) according to the direct sum decomposition. The notation \( {\partial }_{P}f\left( {\cdot ,{x}_{2}}\right) \left( {x}_{1}\right) \) denotes the proximal subdifferential evaluated at \( {x}_{1} \) of the function \( {x}_{1}^{\prime } \mapsto f\left( {{x}_{1}^{\prime },{x}_{2}}\right) \) defined on \( {X}_{1} \) . (This is the partial proximal subdifferential taken with respect to the first coordinate.) Similar considerations apply to \( {\partial }_{P}f\left( {{x}_{1}, \cdot }\right) \left( {x}_{2}\right) \) . The functional analogue to Exercise 1.8 is the following: 2.9. Exercise. Using the notation of the previous paragraph, suppose there exists \( \zeta \in {\partial }_{P}f\left( x\right) \), and write \( \zeta = \left( {{\zeta }_{1},{\zeta }_{2}}\right) \) according to the direct sum decomposition. Show that \( {\zeta }_{1} \in {\partial }_{P}f\left( {\cdot ,{x}_{2}}\right) \left( {x}_{1}\right) \) and \( {\zeta }_{2} \in {\partial }_{P}f\left( {{x}_{1}, \cdot }\right) \left( {x}_{2}\right) \) . Give an example where the converse fails. We will develop proximal calculus in some detail later, but let us note right away that we cannot expect a calculus Sum Rule of the form \[ {\partial }_{P}f\left( x\right) + {\partial }_{P}g\left( x\right) = {\partial }_{P}\left( {f + g}\right) \left( x\right) \] (8) to hold in much generality. One inclusion between these sets can be established easily, but unfortunately, this inclusion is seldom the one we need. 2.10. Exercise. (a) Show that \( {\partial }_{P}f\left( x\right) + {\partial }_{P}g\left( x\right) \subseteq {\partial }_{P}\left( {f + g}\right) \left( x\right) \) . (b) Give an example for which \( {\partial }_{P}\left( {f + g}\right) \left( x\right) \) is nonempty but for which one of \( {\partial }_{P}f\left( x\right) \) or \( {\partial }_{P}g\left( x\right) \) is empty. (c) Show that for all \( c > 0 \), we have \( {\partial }_{P}\left( {cf}\right) \left( x\right) = c{\partial }_{P}f\left( x\right) \) . The following proposition is a proximal Sum Rule, and says in essence that the Sum Rule (8) does hold whenever one of the functions is \( {C}^{2} \) . 2.11. Proposition. Suppose that \( f \in \mathcal{F} \), and let \( x \in X \) . Suppose further that \( g \) is \( {C}^{2} \) in a neighborhood of \( x \) . Then \[ \zeta \in {\partial }_{P}\left( {f + g}\right) \left( x\right) \text{ implies }\zeta - {g}^{\prime }\left( x\right) \in {\partial }_{P}f\left( x\right) . \] Proof. The inequality (3) applied to \( - g \) implies the existence of a constant \( {\sigma }^{\prime } > 0 \) such that \[ - g\left( y\right) + g\left( x\right) + {\sigma }^{\prime }\parallel y - x{\parallel }^{2} \geq \left\langle {-{g}^{\prime }\left( x\right), y - x}\right\rangle \] for all \( y \) near \( x \) . Since \( \zeta \in {\partial }_{P}\left( {f + g}\right) \left( x\right) \), we have \[ f\left( y\right) + g\left( y\right) - f\left( x\right) - g\left( x\right) + \sigma \parallel y - x{\parallel }^{2} \geq \langle \zeta, y - x\rangle \] for some \( \sigma > 0 \) and all \( y \) near \( x \) . Upon adding these inequalities, we arrive at \[ f\left( y\right) - f\left( x\right) + \left( {{\sigma }^{\prime } + \sigma }\right) \parallel y - x{\parallel }^{2} \geq \left\langle {\zeta - {g}^{\prime }\left( x\right), y - x}\right\rangle , \] which holds for all \( y \) near \( x \), and says via Theorem 2.5 that \( \zeta - {g}^{\prime }\left( x\right) \in \) \( {\partial }_{P}f\left( x\right) \) . 2.12. Exercise. Let \( f \in {C}^{2} \), and suppose that \( f \) attains a minimum over \( S \) at \( x \) . Prove that \( - {f}^{\prime }\left( x\right) \in {N}_{S}^{P}\left( x\right) \) . (Hint. Consider \( f + {I}_{S} \) .) A function \( f : X \rightarrow ( - \infty ,\infty \rbrack \) is sai
1012_(GTM178)Nonsmooth Analysis and Control Theory
13
nce \( \zeta \in {\partial }_{P}\left( {f + g}\right) \left( x\right) \), we have \[ f\left( y\right) + g\left( y\right) - f\left( x\right) - g\left( x\right) + \sigma \parallel y - x{\parallel }^{2} \geq \langle \zeta, y - x\rangle \] for some \( \sigma > 0 \) and all \( y \) near \( x \) . Upon adding these inequalities, we arrive at \[ f\left( y\right) - f\left( x\right) + \left( {{\sigma }^{\prime } + \sigma }\right) \parallel y - x{\parallel }^{2} \geq \left\langle {\zeta - {g}^{\prime }\left( x\right), y - x}\right\rangle , \] which holds for all \( y \) near \( x \), and says via Theorem 2.5 that \( \zeta - {g}^{\prime }\left( x\right) \in \) \( {\partial }_{P}f\left( x\right) \) . 2.12. Exercise. Let \( f \in {C}^{2} \), and suppose that \( f \) attains a minimum over \( S \) at \( x \) . Prove that \( - {f}^{\prime }\left( x\right) \in {N}_{S}^{P}\left( x\right) \) . (Hint. Consider \( f + {I}_{S} \) .) A function \( f : X \rightarrow ( - \infty ,\infty \rbrack \) is said to satisfy a Lipschitz condition of rank \( K \) on a given set \( S \) provided that \( f \) is finite on \( S \) and satisfies \[ \left| {f\left( x\right) - f\left( y\right) }\right| \leq K\parallel x - y\parallel \forall x, y \in S. \] 2.13. Exercise. Let \( f \) satisfy a Lipschitz condition of rank \( K \) on some neighborhood of a given point \( {x}_{0} \) . Show that any \( \zeta \in {\partial }_{P}f\left( {x}_{0}\right) \) satisfies \( \parallel \zeta \parallel \leq K \) . The question of a statement converse to that of this exercise is a deeper one that we will address in \( §7 \) . A function \( f \) is said to be Lipschitz near \( x \) if it satisfies the Lipschitz condition (of some rank) on a neighborhood of \( x \) . A function \( f \) is said to be locally Lipschitz on \( S \) if \( f \) is Lipschitz near \( x \) for every \( x \in S \) . 2.14. Exercise. Let \( f \in {C}^{1}\left( U\right), U \) open. Prove that \( f \) is locally Lipschitz on \( U \) . (Hint. Mean Value Theorem.) If \( S \) is a compact convex subset of \( U \), show that \( f \) is Lipschitz on \( S \) with rank \( K \mathrel{\text{:=}} \) \( \max \left\{ {\begin{Vmatrix}{{f}^{\prime }\left( x\right) }\end{Vmatrix} : x \in S}\right\} . \) ## 3 The Density Theorem We now establish an important fact: the set \( \operatorname{dom}\left( {{\partial }_{P}f}\right) \) of points in \( \operatorname{dom}f \) at which at least one proximal subgradient exists is dense in \( \operatorname{dom}f \) . 3.1. Theorem. Suppose \( f \in \mathcal{F} \) . Let \( {x}_{0} \in \operatorname{dom}f \), and let \( \varepsilon > 0 \) be given. Then there exists a point \( y \in {x}_{0} + {\varepsilon B} \) satisfying \( {\partial }_{P}f\left( y\right) \neq \varnothing \) and \( f\left( {x}_{0}\right) - \varepsilon \leq \) \( f\left( y\right) \leq f\left( {x}_{0}\right) \) . In particular, \( \operatorname{dom}\left( {{\partial }_{P}f}\right) \) is dense in \( \operatorname{dom}f \) . Proof. By lower semicontinuity, there exists \( \delta \) with \( 0 < \delta < \varepsilon \) so that \[ x \in {x}_{0} + \delta \bar{B} \Rightarrow f\left( x\right) \geq f\left( {x}_{0}\right) - \varepsilon . \] (1) We first give a simple motivational proof in the case \( X = {\mathbb{R}}^{n} \) . We define \[ g\left( x\right) \mathrel{\text{:=}} \left\{ \begin{array}{ll} {\left\lbrack {\delta }^{2} - {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2}\right\rbrack }^{-1} & \text{ if }\begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix} < \delta , \\ + \infty & \text{ otherwise. } \end{array}\right. \] Note that \( g \) belongs to \( {C}^{2}\left( {{x}_{0} + {\delta B}}\right) \), and that \( g\left( x\right) \rightarrow \infty \) as \( x \) approaches the boundary of \( B\left( {{x}_{0};\delta }\right) \) . Now consider the function \( \left( {f + g}\right) \in \mathcal{F} \) which is bounded below on \( {x}_{0} + \delta \bar{B} \) . It follows that \( f + g \) attains a minimum \( y \) over \( {x}_{0} + \delta \bar{B} \) . Obviously \( y \in {x}_{0} + {\delta B} \), and \( f + g \) has a local minimum at \( y \), thus by Exercise 2.3 we have that \( 0 \in {\partial }_{P}\left( {f + g}\right) \left( y\right) \) . It now follows from Proposition 2.11 that \( - {g}^{\prime }\left( y\right) \in {\partial }_{P}f\left( y\right) \), and in particular that \( {\partial }_{P}f\left( y\right) \neq \varnothing \) . Since \( \delta \) can be made arbitrarily small, it follows that \( \operatorname{dom}\left( {{\partial }_{P}f}\right) \) is dense in \( \operatorname{dom}f \) . In view of (1), we are left only with showing that \( f\left( y\right) \leq f\left( {x}_{0}\right) \) . We deduce this by noting that \( y \) is a minimum of \( f + g \) and \( g\left( {x}_{0}\right) \leq g\left( y\right) \), and hence \[ f\left( y\right) \leq f\left( {x}_{0}\right) + \left( {g\left( {x}_{0}\right) - g\left( y\right) }\right) \leq f\left( {x}_{0}\right) . \] The proof is thus complete if \( X \) is finite dimensional. The proof in the general case is more complicated, due to the possible nonexistence of minimizers. An iterative procedure will lead to an appropriate minimum's existence, however. We begin again by choosing \( \delta \) so that (1) holds, and we take \( \sigma > {2\varepsilon }/{\delta }^{2} \) . We will show that there exists a point \( z \in {x}_{0} + {\delta B} \) so that the function \( x \mapsto \) \( f\left( x\right) + \sigma \parallel x - z{\parallel }^{2} \) has a minimum over \( {x}_{0} + {\delta B} \) at some point \( y \in {x}_{0} + {\delta B} \) that satisfies \( f\left( y\right) \leq f\left( {x}_{0}\right) \) . Once the existence of such \( y \) and \( z \) are established, the proof is easily completed as follows. Since \( y \) is a minimum, we have \( 0 \in {\partial }_{P}\left( {f + \parallel \left( \cdot \right) - z{\parallel }^{2}}\right) \left( y\right) \), which by Proposition 2.11 and Exercise 2.4 gives the inclusion \( - {2\sigma }\left( {y - z}\right) \in {\partial }_{P}f\left( y\right) \) . Also, \( f\left( {x}_{0}\right) - \varepsilon \leq f\left( y\right) \leq f\left( {x}_{0}\right) \) is then immediate in view of (1). We proceed to demonstrate the existence of points \( y \) and \( z \) having the properties described above. We define \[ {S}_{0} \mathrel{\text{:=}} \left\{ {x \in {x}_{0} + \delta \bar{B} : f\left( x\right) + \frac{\sigma }{2}{\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2} \leq f\left( {x}_{0}\right) }\right\} . \] We claim that \( {S}_{0} \) is closed and that \[ {x}_{0} \in {S}_{0} \subset {x}_{0} + {\delta B} \] (2) Indeed, if \( x \in {S}_{0} \), then by (1) and the choice of \( \sigma \), we have \[ {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2} \leq \frac{2}{\sigma }\left\lbrack {f\left( {x}_{0}\right) - f\left( x\right) }\right\rbrack < {\delta }^{2} \] which gives (2). The closedness of \( {S}_{0} \) follows from the lower semicontinuity of \( f \) (see Exercise 2.1(a)). If \( {S}_{0} \) contains only a single point \( y \), then \( y = {x}_{0} \) and the proof is completed, since then \( {x}_{0} \) is a minimum of \[ x \mapsto f\left( x\right) + \sigma {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2} \] over \( x \in {x}_{0} + {\delta B} \) . Since this is generally not the case, we employ an iterative procedure producing successively smaller sets that will shrink to a point \( y \) . Let \( {x}_{1} \in {S}_{0} \) be chosen so that \[ f\left( {x}_{1}\right) + \frac{\sigma }{2}{\begin{Vmatrix}{x}_{1} - {x}_{0}\end{Vmatrix}}^{2} \leq \mathop{\inf }\limits_{{x \in {S}_{0}}}\left\{ {f\left( x\right) + \frac{\sigma }{2}{\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2}}\right\} + \frac{\sigma }{4}, \] and define another closed set by \[ {S}_{1} \mathrel{\text{:=}} \left\{ {x \in {S}_{0} : f\left( x\right) + \sigma \left\lbrack {\frac{{\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2}}{2} + \frac{{\begin{Vmatrix}x - {x}_{1}\end{Vmatrix}}^{2}}{4}}\right\rbrack }\right. \] \[ \left. { \leq f\left( {x}_{1}\right) + \frac{\sigma }{2}{\begin{Vmatrix}{x}_{1} - {x}_{0}\end{Vmatrix}}^{2}}\right\} . \] Note that \( {x}_{1} \in {S}_{1} \), so that \( {S}_{1} \) is nonempty. Again, if \( {x}_{1} \) were the only point in \( {S}_{1} \) the proof would be complete (why?). Inductively, if \( {x}_{j} \) and \( {S}_{j} \) are chosen for \( j \geq 0 \), we choose \( {x}_{j + 1} \in {S}_{j} \) so that \[ f\left( {x}_{j + 1}\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{j}\frac{{\begin{Vmatrix}{x}_{j + 1} - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}} \] \[ \leq \mathop{\inf }\limits_{{s \in {S}_{j}}}\left\{ {f\left( x\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{j}\frac{{\begin{Vmatrix}x - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}}\right\} + \frac{\sigma }{{4}^{j + 1}} \] (3) and we define \[ {S}_{j + 1} \mathrel{\text{:=}} \left\{ {x \in {S}_{j} : f\left( x\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{{j + 1}}\frac{{\begin{Vmatrix}x - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}}\right. \] \[ \left. { \leq f\left( {x}_{j + 1}\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{j}\frac{{\begin{Vmatrix}{x}_{j + 1} - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}}\right\} . \] (4) We obviously have \( {x}_{j + 1} \in {S}_{j + 1} \subset {S}_{j} \) for each \( j \geq 0 \), and thus \( \left\{ {S}_{j}\right\} \) is a nested sequence of nonempty closed sets. We need only show that \( \sup \left\{ {\begin{Vmatrix}{x - {x}^{\prime }}\end{Vmatrix} : x,{x}^{\prime } \in {S}_{j}}\right\} = : \operatorname{diam}\left( {S}_{j}\right) \rightarrow 0 \) as \( j \rightarrow \infty \) to conclude that the sequence \( \left\{ {S}_{j}\right\} \) shrinks to a single point. (This is a fact about complete metric spaces known as Cantor’s Theorem.) Toward this end, let \( x \in {S}_{j + 1} \) . Then for each \( j \geq 0 \), we have by (4) and (3) that \[ \frac{\sigma }{2}\frac{{\begin{Vmatrix}x - {x}_{j + 1}\end{Vmatrix}}^{2}}{{2}^{j + 1}} \leq f\left( {x}_{j + 1}\right)
1012_(GTM178)Nonsmooth Analysis and Control Theory
14
{2}}{{2}^{i}}}\right. \] \[ \left. { \leq f\left( {x}_{j + 1}\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{j}\frac{{\begin{Vmatrix}{x}_{j + 1} - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}}\right\} . \] (4) We obviously have \( {x}_{j + 1} \in {S}_{j + 1} \subset {S}_{j} \) for each \( j \geq 0 \), and thus \( \left\{ {S}_{j}\right\} \) is a nested sequence of nonempty closed sets. We need only show that \( \sup \left\{ {\begin{Vmatrix}{x - {x}^{\prime }}\end{Vmatrix} : x,{x}^{\prime } \in {S}_{j}}\right\} = : \operatorname{diam}\left( {S}_{j}\right) \rightarrow 0 \) as \( j \rightarrow \infty \) to conclude that the sequence \( \left\{ {S}_{j}\right\} \) shrinks to a single point. (This is a fact about complete metric spaces known as Cantor’s Theorem.) Toward this end, let \( x \in {S}_{j + 1} \) . Then for each \( j \geq 0 \), we have by (4) and (3) that \[ \frac{\sigma }{2}\frac{{\begin{Vmatrix}x - {x}_{j + 1}\end{Vmatrix}}^{2}}{{2}^{j + 1}} \leq f\left( {x}_{j + 1}\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{j}\frac{{\begin{Vmatrix}{x}_{j + 1} - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}} \] \[ - \left\{ {f\left( x\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{j}\frac{{\begin{Vmatrix}x - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}}\right\} \leq \frac{\sigma }{{4}^{j + 1}}. \] It follows that \[ \mathop{\sup }\limits_{{x \in {S}_{j + 1}}}\begin{Vmatrix}{x - {x}_{j + 1}}\end{Vmatrix} \leq {2}^{-j/2} \] and thus \( \mathop{\lim }\limits_{{j \rightarrow \infty }}\operatorname{diam}\left( {S}_{j}\right) = 0 \) . Consequently, since the Hilbert space \( X \) is complete, there is a point \( y \) such that \( \mathop{\bigcap }\limits_{{j = 1}}^{\infty }{S}_{j} = \{ y\} \) . Of course, \( y \) lies in \( {S}_{0} \), whence \( \begin{Vmatrix}{y - {x}_{0}}\end{Vmatrix} < \delta < \varepsilon \) (in view of (2)) and \[ f\left( y\right) \leq f\left( y\right) + \frac{\sigma }{2}{\begin{Vmatrix}y - {x}_{0}\end{Vmatrix}}^{2} \leq f\left( {x}_{0}\right) \] (by the definition of \( {S}_{0} \) ). Let \( z \) be the point \( \left( {1/2}\right) \mathop{\sum }\limits_{{j = 0}}^{\infty }{x}_{j}/{2}^{j} \) (the sum is uniformly convergent), and observe that the identity \[ \parallel x - z{\parallel }^{2} = \frac{1}{2}\left( {\mathop{\sum }\limits_{{i = 0}}^{\infty }\frac{{\begin{Vmatrix}x - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}}\right) - c \] (5) holds for any \( x \), where \( c \mathrel{\text{:=}} \left( {1/2}\right) \left( {\mathop{\sum }\limits_{{i = 0}}^{\infty }{\begin{Vmatrix}{x}_{i}\end{Vmatrix}}^{2}/{2}^{i}}\right) - \parallel z{\parallel }^{2} \) . We now show that \( y \) minimizes \( x \mapsto f\left( x\right) + \sigma \parallel x - z{\parallel }^{2} \) over \( {x}_{0} + {\delta B} \) . Note first that since \( {x}_{j + 1} \in {S}_{j} \) for all \( j \), it follows that \[ \left\{ {f\left( {x}_{j}\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{{j - 1}}\frac{{\begin{Vmatrix}{x}_{j} - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}}\right\} \text{ is nonincreasing } \] (6) as \( j \) increases. Now let \( x \in {x}_{0} + {\delta B} \) be different from \( y \), and let \( k \geq 0 \) be the least integer such that \( x \notin {S}_{k} \) . Let \( j \) be any index larger than \( k - 1 \) . Since \( y \in {S}_{j + 1} \), the definition (4) of \( {S}_{j + 1} \) and (6) imply \[ f\left( y\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{{j + 1}}\frac{{\begin{Vmatrix}y - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}} \leq \left\{ {f\left( {x}_{j + 1}\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{j}\frac{{\begin{Vmatrix}{x}_{j + 1} - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}}\right\} \] \[ \leq \left\{ {f\left( {x}_{k}\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{{k - 1}}\frac{{\begin{Vmatrix}{x}_{k} - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}}\right\} . \] (7) Now \( k \) has been chosen so that \( x \in {S}_{k - 1} \smallsetminus {S}_{k} \) (if \( k = 0 \), this should be interpreted with \( {S}_{-1} \mathrel{\text{:=}} {x}_{0} + {\delta B} \) ), and hence the definition of \( {S}_{k} \) yields that \[ f\left( {x}_{k}\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{{k - 1}}\frac{{\begin{Vmatrix}{x}_{k} - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}} < f\left( x\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{k}\frac{{\begin{Vmatrix}x - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}} \] \[ \leq f\left( x\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{\infty }\frac{{\begin{Vmatrix}x - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}. \] (8) We now combine (8) with (7) and let \( j \rightarrow \infty \) to deduce that \[ f\left( y\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{\infty }\frac{{\begin{Vmatrix}y - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}} \leq f\left( x\right) + \frac{\sigma }{2}\mathop{\sum }\limits_{{i = 0}}^{\infty }\frac{{\begin{Vmatrix}x - {x}_{i}\end{Vmatrix}}^{2}}{{2}^{i}}. \] We now can simply add \( - {\sigma c} \) to both sides of this inequality and use the representation (5) to conclude that \[ f\left( y\right) + \sigma \parallel y - z{\parallel }^{2} \leq f\left( x\right) + \sigma \parallel x - z{\parallel }^{2}. \] Since \( x \) was any point in \( {x}_{0} + {\delta B} \), we deduce that \( y \) is a minimum of \( x \mapsto f\left( x\right) + \parallel x - z{\parallel }^{2} \) over \( {x}_{0} + {\delta B} \), as required. ## 4 Minimization Principles As was pointed out at the time, a simple and direct proof of the Density Theorem 3.1 is possible in finite dimensions. The proof fails in infinite dimensions for a reason that is a persistent and thorny issue in analysis: the possible nonexistence of a minimizer. Specifically, a lower semicontinuous function on a closed bounded set may not attain a minimum, or even be bounded below. An important type of result affirms that in certain situations an arbitrarily small perturbation of the function in question will attain a minimum. Here is an example, known as Stegall's minimization principle. 4.1. Theorem. Let \( f \in \mathcal{F} \), and suppose that \( f \) is bounded below on the bounded closed set \( S \subset X \), with \( S \cap \operatorname{dom}f \neq \varnothing \) . Then there exists a dense set of points \( x \) in \( X \) having the property that the function \( y \mapsto f\left( y\right) - \langle x, y\rangle \) attains a unique minimum over \( S \) . The main use of this theorem involves taking \( x \) close to 0, so that a small linear perturbation of \( f \) attains a minimum. The next minimization principle we present is due to Borwein and Preiss, and is more complicated to state. It features a quadratic perturbation and two parameters, and can be used in such a way that the conclusion is strongly related to a pregiven point \( {x}_{0} \) of interest. 4.2. Theorem. Let \( f \in \mathcal{F} \) be bounded below, and let \( \varepsilon > 0 \) . Suppose that \( {x}_{0} \) is a point satisfying \( f\left( {x}_{0}\right) < \mathop{\inf }\limits_{{x \in X}}f\left( x\right) + \varepsilon \) . Then, for any \( \lambda > 0 \) there exist points \( y \) and \( z \) with \[ \begin{Vmatrix}{z - {x}_{0}}\end{Vmatrix} < \lambda ,\;\parallel y - z\parallel < \lambda ,\;f\left( y\right) \leq f\left( {x}_{0}\right) , \] and having the property that the function \[ x \mapsto f\left( x\right) + \frac{\varepsilon }{{\lambda }^{2}}\parallel x - z{\parallel }^{2} \] has a unique minimum at \( x = y \) . The proofs of both these minimization principles appear in the next section, for they are simple consequences of the proximal analysis of "inf-convolutions" (together with the Density Theorem). The following exercise is helpful in understanding the content of Theorem 4.2. 4.3. Exercise. Let \( f,{x}_{0} \), and \( \varepsilon \) be as in the statement of Theorem 4.2, where in addition \( f \) is Fréchet differentiable. Prove that there is a point \( y \) in \( {x}_{0} + 2\sqrt{\varepsilon }B \) such that \( \begin{Vmatrix}{{f}^{\prime }\left( y\right) }\end{Vmatrix} \leq 2\sqrt{\varepsilon } \) and \( f\left( y\right) \leq f\left( {x}_{0}\right) \) . Proceed to deduce the existence of a minimizing sequence \( \left\{ {y}_{i}\right\} \) for \( f \) (i.e., one that satisfies \( \mathop{\lim }\limits_{{i \rightarrow \infty }}f\left( {y}_{i}\right) = \mathop{\inf }\limits_{{x \in X}}f\left( x\right) \) ) such that \( \mathop{\lim }\limits_{{i \rightarrow \infty }}\begin{Vmatrix}{{f}^{\prime }\left( {y}_{i}\right) }\end{Vmatrix} = 0 \) . ## 5 Quadratic Inf-Convolutions The inf-convolution of two functions \( f, g \) is another function \( h \) defined as follows: \[ h\left( x\right) \mathrel{\text{:=}} \mathop{\inf }\limits_{{y \in X}}\{ f\left( y\right) + g\left( {x - y}\right) \} . \] The term "convolution" is suggested by the visual resemblance of this formula to the classical integral convolution formula. Our interest here involves only such inf-convolutions formed between a function \( f \in \mathcal{F} \) and the quadratic function \( x \mapsto \alpha \parallel x{\parallel }^{2} \), where \( \alpha > 0 \) . Such functions have surprisingly far-reaching properties. Given \( f \in \mathcal{F} \) that is bounded below and \( \alpha > 0 \), we define \( {f}_{\alpha } : X \rightarrow \mathbb{R} \) by \[ {f}_{\alpha }\left( x\right) \mathrel{\text{:=}} \mathop{\inf }\limits_{{y \in X}}\left\{ {f\left( y\right) + \alpha \parallel x - y{\parallel }^{2}}\right\} . \] (1) We recall some terminology: \( \left\{ {x}_{i}\right\} \) is said to be a minimizing sequence for an infimum of the type \( \mathop{\inf }\limits_{{x \in S}}g\left( x\right) \) provided that the points \( {x}_{i} \) all lie in \( S \) and satisfy \( \mathop{\lim }\limits_{{i \rightarrow \infty }}g\left( {x}_{i}\right) = \mathop{\inf }\limits_{{x \in S}}g\left( x\right) \) . 5.1. Theorem. Suppose that \( f \in \mathcal{F} \) is bounded below by some constant \( c \), and \( {f}_{\alpha } \) is defined as above with \(
1012_(GTM178)Nonsmooth Analysis and Control Theory
15
F} \) and the quadratic function \( x \mapsto \alpha \parallel x{\parallel }^{2} \), where \( \alpha > 0 \) . Such functions have surprisingly far-reaching properties. Given \( f \in \mathcal{F} \) that is bounded below and \( \alpha > 0 \), we define \( {f}_{\alpha } : X \rightarrow \mathbb{R} \) by \[ {f}_{\alpha }\left( x\right) \mathrel{\text{:=}} \mathop{\inf }\limits_{{y \in X}}\left\{ {f\left( y\right) + \alpha \parallel x - y{\parallel }^{2}}\right\} . \] (1) We recall some terminology: \( \left\{ {x}_{i}\right\} \) is said to be a minimizing sequence for an infimum of the type \( \mathop{\inf }\limits_{{x \in S}}g\left( x\right) \) provided that the points \( {x}_{i} \) all lie in \( S \) and satisfy \( \mathop{\lim }\limits_{{i \rightarrow \infty }}g\left( {x}_{i}\right) = \mathop{\inf }\limits_{{x \in S}}g\left( x\right) \) . 5.1. Theorem. Suppose that \( f \in \mathcal{F} \) is bounded below by some constant \( c \), and \( {f}_{\alpha } \) is defined as above with \( \alpha > 0 \) . Then \( {f}_{\alpha } \) is bounded below by \( c \) , and is Lipschitz on each bounded subset of \( X \) (and in particular is finite-valued). Furthermore, suppose \( x \in X \) is such that \( {\partial }_{P}{f}_{\alpha }\left( x\right) \) is nonempty. Then there exists a point \( \bar{y} \in X \) satisfying the following: (a) If \( \left\{ {y}_{i}\right\} \subset X \) is a minimizing sequence for the infimum in (1), then \( \mathop{\lim }\limits_{{i \rightarrow \infty }}{y}_{i} = \bar{y}. \) (b) The infimum in (1) is attained uniquely at \( \bar{y} \) . (c) The Fréchet derivative \( {f}_{a}^{\prime }\left( x\right) \) exists and equals \( {2\alpha }\left( {x - \bar{y}}\right) \) . Thus the proximal subgradient \( {\partial }_{P}{f}_{\alpha }\left( x\right) \) is the singleton \( \{ {2\alpha }\left( {x - \bar{y}}\right) \} \) . (d) \( {2\alpha }\left( {x - \bar{y}}\right) \in {\partial }_{P}f\left( \bar{y}\right) \) . Proof. Suppose we are given \( f \) and \( \alpha > 0 \) as above. It is clear from the definition that \( {f}_{\alpha } \) is bounded below by \( c \) . We now show that \( {f}_{\alpha } \) is Lipschitz on any bounded set \( S \subset X \) . For any fixed \( {x}_{0} \in \operatorname{dom}f \neq \varnothing \), note that \( {f}_{\alpha }\left( x\right) \leq f\left( {x}_{0}\right) + \alpha {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2} \) for all \( x \in X \), and thus in particular \( m \mathrel{\text{:=}} \sup \left\{ {{f}_{\alpha }\left( x\right) : x \in S}\right\} < \infty \) . Since \( \alpha > 0 \) , and \( f \) is bounded below, we have that for any \( \varepsilon > 0 \) the set \[ C \mathrel{\text{:=}} \left\{ {z : \exists y \in S\text{ so that }f\left( z\right) + \alpha \parallel y - z{\parallel }^{2} \leq m + \varepsilon }\right\} \] is bounded in \( X \) . Now let \( x \) and \( y \) belong to \( S \) and \( \varepsilon > 0 \) . Since \( {f}_{\alpha }\left( \cdot \right) \) is given as an infimum, there exists \( z \in C \) so that \[ {f}_{\alpha }\left( y\right) \geq f\left( z\right) + \alpha \parallel y - z{\parallel }^{2} - \varepsilon . \] Thus we have \[ {f}_{\alpha }\left( x\right) - {f}_{\alpha }\left( y\right) \leq {f}_{\alpha }\left( x\right) - f\left( z\right) - \alpha \parallel y - z{\parallel }^{2} + \varepsilon \] \[ \leq f\left( z\right) + \alpha \parallel x - z{\parallel }^{2} - f\left( z\right) - \alpha \parallel y - z{\parallel }^{2} + \varepsilon \] \[ = \alpha \parallel x - y{\parallel }^{2} - {2\alpha }\langle x - y, z - y\rangle + \varepsilon \] \[ \leq \lambda \parallel x - y\parallel + \varepsilon \] where \( \lambda \mathrel{\text{:=}} \alpha \sup \left\{ {\begin{Vmatrix}{{s}^{\prime } - s}\end{Vmatrix} + 2\parallel z - s\parallel : {s}^{\prime }, s \in S, z \in C}\right\} < \infty \) . Reversing the roles of \( x \) and \( y \), and then letting \( \varepsilon \downarrow 0 \), the above shows that \( {f}_{\alpha } \) is Lipschitz of rank \( \lambda \) on \( S \) . We now consider the other assertions in the theorem. Suppose \( x \in X \) is such that there exists at least one \( \zeta \in {\partial }_{P}{f}_{\alpha }\left( x\right) \) . By the proximal subgradient inequality, there exist positive constants \( \sigma \) and \( \eta \) so that \[ \langle \zeta, y - x\rangle \leq {f}_{\alpha }\left( y\right) - {f}_{\alpha }\left( x\right) + \sigma \parallel y - x{\parallel }^{2} \] (2) for all \( y \in B\left( {x;\eta }\right) \) . Now suppose \( \left\{ {y}_{i}\right\} \) is any minimizing sequence of (1) and thus there exists \( {\varepsilon }_{i} \) with \( {\varepsilon }_{i} \downarrow 0 \) as \( i \rightarrow \infty \) such that \[ {f}_{\alpha }\left( x\right) \leq f\left( {y}_{i}\right) + \alpha {\begin{Vmatrix}{y}_{i} - x\end{Vmatrix}}^{2} = {f}_{\alpha }\left( x\right) + {\varepsilon }_{i}^{2}. \] (3) We observe that \[ {f}_{\alpha }\left( y\right) \leq f\left( {y}_{i}\right) + \alpha {\begin{Vmatrix}{y}_{i} - y\end{Vmatrix}}^{2} \] (4) since \( {f}_{\alpha } \) is defined as an infimum over \( X \) . Inserting the inequalities (3) and (4) into (2) yields for each \( y \in B\left( {x;\eta }\right) \) the conclusion \[ \langle \zeta, y - x\rangle \leq \alpha {\begin{Vmatrix}{y}_{i} - y\end{Vmatrix}}^{2} - \alpha {\begin{Vmatrix}{y}_{i} - x\end{Vmatrix}}^{2} + {\varepsilon }_{i}^{2} + \sigma \parallel y - x{\parallel }^{2} \] \[ = - {2\alpha }\langle {y}_{i}, y - x\rangle + \alpha \parallel y{\parallel }^{2} - \alpha \parallel x{\parallel }^{2} + {\varepsilon }_{i}^{2} + \sigma \parallel y - x{\parallel }^{2} \] \[ = {2\alpha }\left\langle {x - {y}_{i}, y - x}\right\rangle + {\varepsilon }_{i}^{2} + \left( {\alpha + \sigma }\right) \parallel y - x{\parallel }^{2}, \] which rewritten says that \[ \left\langle {\zeta - {2\alpha }\left( {x - {y}_{i}}\right), y - x}\right\rangle \leq {\varepsilon }_{i}^{2} + \left( {\alpha + \sigma }\right) \parallel y - x{\parallel }^{2} \] (5) for all \( y \in B\left( {x;\eta }\right) \) . Now let \( v \in B \) . Note that \( y \mathrel{\text{:=}} x + {\varepsilon }_{i}v \in B\left( {x;\eta }\right) \) for large \( i \) since \( {\varepsilon }_{i} \downarrow 0 \) as \( i \rightarrow \infty \) . Hence for all large \( i \), we can insert this value of \( y \) into (5) to deduce \[ \left\langle {\zeta - {2\alpha }\left( {x - {y}_{i}}\right), v}\right\rangle \leq {\varepsilon }_{i}\left( {1 + \alpha + \sigma }\right) \] Since \( v \in B \) is arbitrary, it follows that \[ \begin{Vmatrix}{\zeta - {2\alpha }\left( {x - {y}_{i}}\right) }\end{Vmatrix} \leq {\varepsilon }_{i}\left( {1 + \alpha + \sigma }\right) \] (6) We now define \( \bar{y} \in X \) by \[ \bar{y} \mathrel{\text{:=}} x - \frac{\zeta }{2\alpha } \] and observe that (a) immediately follows from letting \( i \rightarrow \infty \) in (6). To see that \( \bar{y} \) achieves the infimum in (1), it suffices to observe from (a) and the lower semicontinuity of \( f \) that \[ {f}_{\alpha }\left( x\right) \leq f\left( \bar{y}\right) + \alpha \parallel \bar{y} - x{\parallel }^{2} \leq \mathop{\liminf }\limits_{{i \rightarrow \infty }}\left\lbrack {f\left( {y}_{i}\right) + \alpha {\begin{Vmatrix}{y}_{i} - x\end{Vmatrix}}^{2}}\right\rbrack = {f}_{\alpha }\left( x\right) , \] where the last equality comes from (3). It is also clear that \( \bar{y} \) is unique, since if \( \widehat{y} \) is another minimizer of (1), the constant sequence \( {y}_{i} \mathrel{\text{:=}} \widehat{y} \) is minimizing, and therefore must converge to \( \bar{y} \) by (a). Hence (b) holds. The following observation about a supergradient will be useful in proving the Fréchet differentiability assertion: \[ {2\alpha }\left( {x - \bar{y}}\right) \in {\partial }^{P}{f}_{\alpha }\left( x\right) \] (7) To see this, let \( y \in X \) and observe that \[ {f}_{\alpha }\left( y\right) \leq f\left( \bar{y}\right) + \alpha \parallel y - \bar{y}{\parallel }^{2} \] with equality holding if \( y = x \) . Then we see that \[ {f}_{\alpha }\left( x\right) - {f}_{\alpha }\left( y\right) \geq f\left( \bar{y}\right) + \alpha \parallel x - \bar{y}{\parallel }^{2} - f\left( \bar{y}\right) - \alpha \parallel y - \bar{y}{\parallel }^{2} \] \[ = \langle {2\alpha }\left( {x - \bar{y}}\right), y - x\rangle - \alpha \parallel x - y{\parallel }^{2}. \] This confirms (7) (see Exercise 2.8(a)). Part (c) of the theorem now follows from (7) combined with Exercise 2.8(b). As for part (d), observe that the function \( {x}^{\prime } \mapsto f\left( {x}^{\prime }\right) + \alpha {\begin{Vmatrix}{x}^{\prime } - x\end{Vmatrix}}^{2} \) attains a minimum at \( {x}^{\prime } = \bar{y} \), so that its proximal subdifferential there contains 0 . With the help of Proposition 2.11 and Exercise 2.4(a), this translates to precisely statement (d). Two immediate consequences of Theorem 5.1 are the minimization principles presented in \( §4 \), as we now see. ## 5.2. Corollary. Theorem 4.1 holds. Proof. Suppose \( S \subset X \) is nonempty, closed, and bounded, and \( f \in \mathcal{F} \) is bounded below on \( S \) with \( \operatorname{dom}f \cap S \neq \varnothing \) . Define \[ g\left( x\right) \mathrel{\text{:=}} \mathop{\inf }\limits_{{y \in X}}\left\{ {f\left( y\right) + {I}_{S}\left( y\right) - \frac{1}{2}\parallel y{\parallel }^{2} + \frac{1}{2}\parallel x - y{\parallel }^{2}}\right\} \] (8) which is easily seen to be a function of the form \( {f}_{\alpha } \) as in (1) (where \( f = \) \( f + {I}_{S} - \left( {1/2}\right) \parallel \cdot {\parallel }^{2} \) and \( \alpha = 1/2) \) . Furthermore, expression (8) for \( g\left( x\right) \) can be simplified to \[ g\left( x\right) = \mathop{\inf }\limits_{{y \in S}}\{ f\left( y\right) - \langle x, y\rangle \} + \frac{1}{2}\parallel x{\parallel }^{2}. \] (9) It is clear that for fixed \( x \in X \), the sets of \( y \) attaining the infima in (8), (9), and in \[ \bar{g}\left( x\right) \mathrel{\text{:=}} \mathop{\inf }\limits_{{y \in S}
1012_(GTM178)Nonsmooth Analysis and Control Theory
16
set X \) is nonempty, closed, and bounded, and \( f \in \mathcal{F} \) is bounded below on \( S \) with \( \operatorname{dom}f \cap S \neq \varnothing \) . Define \[ g\left( x\right) \mathrel{\text{:=}} \mathop{\inf }\limits_{{y \in X}}\left\{ {f\left( y\right) + {I}_{S}\left( y\right) - \frac{1}{2}\parallel y{\parallel }^{2} + \frac{1}{2}\parallel x - y{\parallel }^{2}}\right\} \] (8) which is easily seen to be a function of the form \( {f}_{\alpha } \) as in (1) (where \( f = \) \( f + {I}_{S} - \left( {1/2}\right) \parallel \cdot {\parallel }^{2} \) and \( \alpha = 1/2) \) . Furthermore, expression (8) for \( g\left( x\right) \) can be simplified to \[ g\left( x\right) = \mathop{\inf }\limits_{{y \in S}}\{ f\left( y\right) - \langle x, y\rangle \} + \frac{1}{2}\parallel x{\parallel }^{2}. \] (9) It is clear that for fixed \( x \in X \), the sets of \( y \) attaining the infima in (8), (9), and in \[ \bar{g}\left( x\right) \mathrel{\text{:=}} \mathop{\inf }\limits_{{y \in S}}\{ f\left( y\right) - \langle x, y\rangle \} \] (10) all coincide. The Density Theorem 3.1 says that \( \operatorname{dom}{\partial }_{P}g \) is dense in \( \operatorname{dom}g = \) \( X \), and Theorem 5.1(b) says that for each \( x \in \operatorname{dom}{\partial }_{P}g \), the infimum in (8) is uniquely attained. Hence for a dense set of \( x \in X \), the infimum in (10) is attained at a unique \( y \in S \), which is the assertion of the theorem. ## 5.3. Corollary. Theorem 4.2 holds. Proof. Let \( f,{x}_{0} \), and \( \varepsilon \) be as in the statement of Theorem 4.2, and let \( \lambda > 0 \) . Consider the function \( {f}_{\alpha } \) as given in (1) with \( \alpha = \varepsilon /{\lambda }^{2} \) : \[ {f}_{\alpha }\left( x\right) \mathrel{\text{:=}} \mathop{\inf }\limits_{{{y}^{\prime } \in X}}\left\{ {f\left( {y}^{\prime }\right) + \frac{\varepsilon }{{\lambda }^{2}}{\begin{Vmatrix}{y}^{\prime } - x\end{Vmatrix}}^{2}}\right\} . \] By the Density Theorem 3.1, there exists \( z \in {x}_{0} + {\lambda B} \) satisfying \( {f}_{\alpha }\left( z\right) \leq \) \( {f}_{\alpha }\left( {x}_{0}\right) \leq f\left( {x}_{0}\right) \) and with \( {\partial }_{P}{f}_{\alpha }\left( z\right) \neq \varnothing \) . By Theorem 5.1(b), there is a unique point \( y \) at which the infimum defining \( {f}_{\alpha }\left( z\right) \) is attained. All the assertions of the theorem are now immediate, except for the inequality \( \parallel y - z\parallel < \lambda \) , which we proceed to confirm. We have \[ f\left( y\right) + \frac{\varepsilon }{{\lambda }^{2}}\parallel y - z{\parallel }^{2} = {f}_{\alpha }\left( z\right) \leq f\left( {x}_{0}\right) < \mathop{\inf }\limits_{X}\left( f\right) + \varepsilon , \] and so \[ \frac{\varepsilon }{{\lambda }^{2}}\parallel y - z{\parallel }^{2} < \mathop{\inf }\limits_{X}\left( f\right) - f\left( y\right) + \varepsilon \leq \varepsilon \] which implies the inequality we seek. 5.4. Remark. The proof of Theorem 4.2 shows that we may in fact take \( z \) arbitrarily near \( {x}_{0} \) (and not merely within distance \( \lambda \) of \( {x}_{0} \) ). ## 6 The Distance Function In this section we examine the proximal subgradients of the distance function \( {d}_{S} \) associated to a nonempty closed subset \( S \) of \( X \) . The results of the analysis will allow us to deduce geometric analogues of the minimization principles of \( §4 \) . 6.1. Theorem. Suppose \( x \notin S \) and \( \zeta \in {\partial }_{P}{d}_{S}\left( x\right) \) . Then there exists a point \( \bar{s} \in S \) so that the following holds: (a) Every minimizing sequence \( \left\{ {s}_{i}\right\} \subset S \) of \( \mathop{\inf }\limits_{{s \in S}}\parallel s - x\parallel \) converges to \( \bar{s} \) . (b) The set of closest points \( {\operatorname{proj}}_{S}\left( x\right) \) in \( S \) to \( x \) is the singleton \( \{ \bar{s}\} \) . (c) The Fréchet derivative \( {d}_{S}^{\prime }\left( x\right) \) exists, and \[ \{ \zeta \} = {\partial }_{P}{d}_{S}\left( x\right) = \left\{ {{d}_{S}^{\prime }\left( x\right) }\right\} = \left\{ \frac{x - \bar{s}}{\parallel x - \bar{s}\parallel }\right\} . \] (d) \( \zeta \in {N}_{S}^{P}\left( \bar{s}\right) \) . Proof. Suppose \( x \notin S \) and \( \zeta \in {\partial }_{P}{d}_{S}\left( x\right) \) . There exists \( \sigma > 0 \) and \( \eta > 0 \) so that \[ {d}_{S}\left( y\right) - {d}_{S}\left( x\right) \geq \langle \zeta, y - x\rangle - \sigma \parallel y - x{\parallel }^{2}\forall y \in B\left( {x;\eta }\right) . \] (1) Note that for all \( y \in X \) we have \[ {d}_{S}^{2}\left( y\right) = \mathop{\inf }\limits_{{z \in X}}\left\{ {{I}_{S}\left( z\right) + \parallel z - y{\parallel }^{2}}\right\} \] (2) Hence for the choices \( \alpha = 1 \) and \( f = {I}_{S} \), we see that \( {d}_{S}^{2}\left( \cdot \right) \) is a quadratic inf-convolution function. The assertions of the present theorem will be derived as consequences of Theorem 5.1, but in order to do so, we must first establish that subgradients of \( {d}_{S} \) lead to subgradients of \( {d}_{S}^{2} \) . In particular, we show that \( \zeta \in {\partial }_{P}{d}_{S}\left( x\right) \) implies \( 2{d}_{S}\left( x\right) \zeta \in \partial {d}_{S}^{2}\left( x\right) \), which can be viewed as a special case of the Chain Rule. Observe that the elementary identity \[ {d}_{S}^{2}\left( y\right) - {d}_{S}^{2}\left( x\right) = 2{d}_{S}\left( x\right) \left( {{d}_{S}\left( y\right) - {d}_{S}\left( x\right) }\right) + {\left( {d}_{S}\left( y\right) - {d}_{S}\left( x\right) \right) }^{2} \] holds for any \( y \in X \) . Using it and (1) we see for each \( y \in B\left( {x;\eta }\right) \) that \[ {d}_{S}^{2}\left( y\right) - {d}_{S}^{2}\left( x\right) \geq 2{d}_{S}\left( x\right) \left( {{d}_{S}\left( y\right) - {d}_{S}\left( x\right) }\right) \] \[ \geq \left\langle {2{d}_{S}\left( x\right) \zeta, y - x}\right\rangle - 2{d}_{S}\left( x\right) \sigma \parallel y - x{\parallel }^{2}, \] which implies \( 2{d}_{S}\left( x\right) \zeta \in {\partial }_{P}{d}_{S}^{2}\left( x\right) \), as claimed. Hence parts (a) and (b) of the present theorem follow from parts (a) and (b) of Theorem 5.1, respectively, since minimizing sequences and minima of (2) are precisely those for the infimum defining \( {d}_{S}\left( x\right) \) . We also have from Theorem 5.1(c) that \( {\partial }_{P}{d}_{S}\left( x\right) = \{ \zeta \} \) and \[ \zeta = \frac{x - \bar{s}}{{d}_{S}\left( x\right) } = \frac{x - \bar{s}}{\parallel x - \bar{s}\parallel }. \] We turn now to differentiability. Theorem 5.1 implies that \( {d}_{S}^{2}\left( \cdot \right) \) is Fréchet differentiable at \( x \), whence for any \( v \in X \), the limit as \( t \downarrow 0 \) of \[ \frac{{d}_{S}^{2}\left( {x + {tv}}\right) - {d}_{S}^{2}\left( x\right) }{t} = \left\lbrack \frac{{d}_{S}\left( {x + {tv}}\right) - {d}_{S}\left( x\right) }{t}\right\rbrack \left\lbrack {{d}_{S}\left( {x + {tv}}\right) + {d}_{S}\left( x\right) }\right\rbrack \] exists, with the convergence being uniform for \( v \) in bounded sets. We also know that the limit is \( 2{d}_{S}\left( x\right) \langle \zeta, v\rangle \) . Since \( {d}_{S}\left( {x + {tv}}\right) + {d}_{S}\left( x\right) \) evidently converges to \( 2{d}_{S}\left( x\right) > 0 \) uniformly for \( v \) in bounded sets as \( t \downarrow 0 \) (as a consequence of the fact that \( {d}_{S} \) is globally Lipschitz), it follows that the difference quotient \[ \frac{{d}_{S}\left( {x + {tv}}\right) - {d}_{S}\left( x\right) }{t} \] converges to \( \langle \zeta, v\rangle \) uniformly on bounded sets as \( t \downarrow 0 \) ; i.e., \( {d}_{S}^{\prime }\left( x\right) = \zeta \) . Finally, we have that \( \bar{s} \) minimizes the function \( s \mapsto \parallel s - x{\parallel }^{2} \) over \( s \in \) \( S \) . It follows from Exercise 2.12 that \( 2\left( {x - \bar{s}}\right) \in {N}_{S}^{P}\left( \bar{s}\right) \), which implies \( \zeta = \left( {x - \bar{s}}\right) /\parallel x - \bar{s}\parallel \in {N}_{S}^{P}\left( \bar{s}\right) \), since this set is a cone. The theorem is proved. The following corollary asserts that the points that admit unique closest points to a given closed set are dense (cf. Exercise 1.2(d)). 6.2. Corollary. Suppose \( S \subset X \) is closed. (a) There is a dense set of points in \( X \smallsetminus S \) which admit unique closest points in \( S \) . (b) The set of points \( s \in \operatorname{bdry}S \) for which \( {N}_{S}^{P}\left( s\right) \neq \{ 0\} \) is dense in bdry \( S \) . Proof. By the Density Theorem 3.1, \( \operatorname{dom}{\partial }_{P}{d}_{S}\left( \cdot \right) \) is dense in \( X \), and by Theorem 6.1 each point \( x \) with \( {\partial }_{P}{d}_{S}\left( x\right) \neq \varnothing \) has a unique closest point in \( S \) . This proves (a). To see that (b) holds, let \( s \in \operatorname{bdry}S \) and \( \varepsilon > 0 \) . By (a), there exists \( x \notin S \) with \( \parallel x - s\parallel < \varepsilon \) and so that \( {\operatorname{proj}}_{S}\left( x\right) = \{ \bar{s}\} \) . Then \( {N}_{S}^{P}\left( \bar{s}\right) \) contains \( x - \bar{s} \neq 0 \) . Since we have \[ \parallel \bar{s} - s\parallel \leq \parallel x - \bar{s}\parallel + \parallel x - s\parallel \leq 2\parallel x - s\parallel \leq {2\varepsilon } \] assertion (b) follows. The next proposition illustrates a mechanism for using the distance function as a tool in solving constrained optimization problems. Such a problem is of the form \[ \inf f\left( s\right) \text{subject to}s \in S\text{,} \] (3) where \( f \in \mathcal{F} \) and \( S \subset X \) is closed. We will study constrained optimization problems in detail in Chapter 3. The technique introduced below is called exact penalization, since it adds a penalty to the function to be minimized so as to obtain an equivalence between the original constrained problem and the new penalized and unconstrained problem. Exact penalization has
1012_(GTM178)Nonsmooth Analysis and Control Theory
17
parallel < \varepsilon \) and so that \( {\operatorname{proj}}_{S}\left( x\right) = \{ \bar{s}\} \) . Then \( {N}_{S}^{P}\left( \bar{s}\right) \) contains \( x - \bar{s} \neq 0 \) . Since we have \[ \parallel \bar{s} - s\parallel \leq \parallel x - \bar{s}\parallel + \parallel x - s\parallel \leq 2\parallel x - s\parallel \leq {2\varepsilon } \] assertion (b) follows. The next proposition illustrates a mechanism for using the distance function as a tool in solving constrained optimization problems. Such a problem is of the form \[ \inf f\left( s\right) \text{subject to}s \in S\text{,} \] (3) where \( f \in \mathcal{F} \) and \( S \subset X \) is closed. We will study constrained optimization problems in detail in Chapter 3. The technique introduced below is called exact penalization, since it adds a penalty to the function to be minimized so as to obtain an equivalence between the original constrained problem and the new penalized and unconstrained problem. Exact penalization has many theoretical and numerical applications in optimization. 6.3. Proposition. Suppose \( S \) is a closed subset of \( X \), and that \( f \) is Lipschitz of rank \( K \) on an open set \( U \) that contains \( S \) . Assume that \( s \in S \) solves (3). Then the function \( x \mapsto f\left( x\right) + K{d}_{S}\left( x\right) \) attains its minimum over \( U \) at \( x = s \) . Conversely, if \( {K}^{\prime } > K \) and \( x \mapsto f\left( x\right) + {K}^{\prime }{d}_{S}\left( x\right) \) attains its minimum over \( U \) at \( x = s \), then \( s \) belongs to \( S \) and solves (3). Proof. Suppose \( x \in U \) and \( \varepsilon > 0 \) . Let \( {s}^{\prime } \in S \) be such that \( \begin{Vmatrix}{x - {s}^{\prime }}\end{Vmatrix} \leq \) \( {d}_{S}\left( x\right) + \varepsilon \) . Since \( f \) is minimized over \( S \) at \( s \), and using the Lipschitz property, we have \[ f\left( s\right) \leq f\left( {s}^{\prime }\right) \] \[ \leq f\left( x\right) + K\begin{Vmatrix}{{s}^{\prime } - x}\end{Vmatrix} \] \[ \leq f\left( x\right) + K{d}_{S}\left( x\right) + {\varepsilon K} \] Letting \( \varepsilon \downarrow 0 \) shows that \( f + K{d}_{S} \) attains its minimum over \( U \) at \( x = s \) . To prove the converse, suppose \( {K}^{\prime } > K \) and that a point \( s \notin S \) minimizes \( f + {K}^{\prime }{d}_{S} \) over \( U \) . Pick \( {s}^{\prime } \in S \) so that \( \begin{Vmatrix}{{s}^{\prime } - s}\end{Vmatrix} < \left( {{K}^{\prime }/K}\right) {d}_{S}\left( s\right) \) . Since \( f \) is Lipschitz of rank \( K \), we have \[ f\left( {s}^{\prime }\right) \leq f\left( s\right) + K\begin{Vmatrix}{{s}^{\prime } - s}\end{Vmatrix}. \] We are assuming that \( s \) minimizes \( x \mapsto f\left( x\right) + {K}^{\prime }{d}_{S}\left( x\right) \) over \( U \), and since \( {s}^{\prime } \in S \subset U \), we conclude that \[ f\left( s\right) + {K}^{\prime }{d}_{S}\left( s\right) \leq f\left( {s}^{\prime }\right) \leq f\left( s\right) + K\begin{Vmatrix}{{s}^{\prime } - s}\end{Vmatrix} < f\left( s\right) + {K}^{\prime }{d}_{S}\left( s\right) , \] which is a contradiction. Thus \( s \in S \) . That \( s \) solves (3) is then immediate. The next result forges another link between the geometric and functional points of view. 6.4. Theorem. Suppose \( S \) is closed and \( s \in S \) . Then \[ {N}_{S}^{P}\left( s\right) = \left\{ {{t\zeta } : t \geq 0,\zeta \in {\partial }_{P}{d}_{S}\left( s\right) }\right\} . \] Proof. Suppose \( \zeta \in {N}_{S}^{P}\left( s\right) \) . By the proximal normal inequality, there exists \( \sigma > 0 \) so that \[ \left\langle {\zeta ,{s}^{\prime } - s}\right\rangle \leq \sigma {\begin{Vmatrix}{s}^{\prime } - s\end{Vmatrix}}^{2} \] (4) for all \( {s}^{\prime } \in S \) . It immediately follows that the \( {C}^{2} \) function \( x \mapsto - \langle \zeta, x\rangle + \) \( \sigma \parallel x - s{\parallel }^{2} \) has a minimum over \( S \) at \( x = s \) . Since this function is locally Lipschitz, we conclude from Proposition 6.3 that for some constant \( K \), the function \[ x \mapsto - \langle \zeta, x\rangle + \sigma \parallel x - s{\parallel }^{2} + K{d}_{S}\left( x\right) \] has a local minimum at \( x = s \) . It follows that \( \zeta /K \in {\partial }_{P}{d}_{S}\left( s\right) \), which proves one of the desired inclusions. To prove the reverse inclusion, suppose \( \zeta \in {\partial }_{P}{d}_{S}\left( s\right) \) . The proximal subgra-dient inequality provides a constant \( \sigma \) so that \[ {d}_{S}\left( x\right) - \langle \zeta, x - s\rangle + \sigma \parallel x - s{\parallel }^{2} \geq {d}_{S}\left( s\right) = 0 \] for all \( x \) near \( s \) . In particular, there exists \( \delta > 0 \) so that (4) holds for all \( {s}^{\prime } \in S \cap \{ s + {\delta B}\} \), which is equivalent to \( \zeta \in {N}_{S}^{P}\left( s\right) \) by Proposition 1.5(b). Since \( {N}_{S}^{P}\left( s\right) \) is a cone, the proof is now complete. ## 7 Lipschitz Functions Among the functions that have the Lipschitz property are the distance function \( {d}_{S} \), the quadratic inf-convolution functions of \( §5,{C}^{1} \) functions and, as we now see, all finite (nonpathological) convex functions. A function \( f \) is said to be bounded above near \( x \) if there exists \( \eta > 0 \) and \( r \in \mathbb{R} \) such that \[ f\left( y\right) \leq r\forall y \in B\left( {x;\eta }\right) . \] 7.1. Proposition. Let \( U \subset X \) be open and convex, and let \( f : U \rightarrow \mathbb{R} \) be convex and finite on \( U \) . Suppose that \( f \) is bounded above at some point \( \bar{x} \in U \) . Then \( f \) is locally Lipschitz on \( U \) . Proof. Let \( x \in U \), and let us suppose for the moment that \( f \) is bounded on a neighborhood of \( x \) . Choose \( \eta > 0 \) and \( r \in \mathbb{R} \) so that \( y \in B\left( {x;{2\eta }}\right) \subset U \) implies that \( \left| {f\left( y\right) }\right| \leq r \) . Let \( {y}_{1},{y}_{2} \) be distinct points in \( B\left( {x;\eta }\right) \), and set \( \delta = \begin{Vmatrix}{{y}_{1} - {y}_{2}}\end{Vmatrix} \) . Let \( {y}_{3} = {y}_{2} + \left( {\eta /\delta }\right) \left( {{y}_{2} - {y}_{1}}\right) \in B\left( {x;{2\eta }}\right) \) and note then that \[ {y}_{2} = \frac{\eta }{\eta + \delta }{y}_{1} + \frac{\delta }{\eta + \delta }{y}_{3} \] By convexity, we have that \[ f\left( {y}_{2}\right) \leq \frac{\eta }{\eta + \delta }f\left( {y}_{1}\right) + \frac{\delta }{\eta + \delta }f\left( {y}_{3}\right) \] which can be rearranged to \[ f\left( {y}_{2}\right) - f\left( {y}_{1}\right) \leq \frac{\delta }{\eta + \delta }\left\lbrack {f\left( {y}_{3}\right) - f\left( {y}_{1}\right) }\right\rbrack \leq \frac{\delta }{\eta }\left| {f\left( {y}_{3}\right) - f\left( {y}_{1}\right) }\right| . \] Since \( \left| {f\left( y\right) }\right| \leq r \) on \( B\left( {x;{2\eta }}\right) \) and \( \delta = \begin{Vmatrix}{{y}_{1} - {y}_{2}}\end{Vmatrix} \), we conclude that \[ f\left( {y}_{2}\right) - f\left( {y}_{1}\right) \leq \frac{2r}{\eta }\begin{Vmatrix}{{y}_{1} - {y}_{2}}\end{Vmatrix} \] Switching the roles of \( {y}_{1} \) and \( {y}_{2} \) yields that \( f \) is Lipschitz on \( B\left( {x;\eta }\right) \) . We now show \( f \) is locally bounded near any \( x \in U \) . Without loss of generality, we assume \( \bar{x} = 0 \) . Suppose that \( f\left( y\right) \leq r \) for all \( y \) satisfying \( \parallel y\parallel < \eta \) , where \( B\left( {0;\eta }\right) \subset U \) ; let \( x \in U \) . There exists \( t \in \left( {0,1}\right) \) so that \( z \mathrel{\text{:=}} x/t \in U \) , and we claim that \( f \) is bounded above on \( B\left( {x;\left( {1 - t}\right) \eta }\right) \) . Indeed, let \( v \in B\left( {x;\left( {1 - t}\right) \eta }\right) \), and choose \( y \in {\eta B} \) so that \( v = x + \left( {1 - t}\right) y = {tz} + \left( {1 - t}\right) y \) . By convexity, we have \[ f\left( v\right) \leq {tf}\left( z\right) + \left( {1 - t}\right) f\left( y\right) \leq {tf}\left( z\right) + r \] which shows that \( f \) is bounded above on \( B\left( {x;\left( {1 - t}\right) \eta }\right) \) . We complete the proof by showing that \( f \) is bounded below on the same ball. Let \( \bar{r} \) be an upper bound for \( f \) on \( B\left( {x;\left( {1 - t}\right) \eta }\right) \), and pick any point \( u \) in this set. Choose \( {u}^{\prime } \in B\left( {x;\left( {1 - t}\right) \eta }\right) \) so that \( \left( {u + {u}^{\prime }}\right) /2 = x \) . Then \( f\left( x\right) \leq \left( {f\left( u\right) + f\left( {u}^{\prime }\right) }\right) /2 \) implies \( f\left( u\right) \geq {2f}\left( x\right) - f\left( {u}^{\prime }\right) \geq {2f}\left( x\right) - \bar{r} \), establishing the required lower bound. 7.2. Exercise. Show that the boundedness hypothesis of Proposition 7.1 is automatically satisfied if \( X = {\mathbb{R}}^{n} \) . (Hint. There exist finitely many points \( {p}_{i} \) such that \( x \in \operatorname{int}\operatorname{co}\left\{ {p}_{i}\right\} \subset U \) .) The class of Lipschitz functions constitutes an extremely interesting one in its own right. We now show in the general setting that the Lipschitz property can be characterized in proximal terms. It is interesting to compare the nature of this proof to the very different proofs in the smooth case, which use integration or the Mean Value Theorem. Here, optimization is the principal tool, and the criterion need only hold at certain points (where \( {\partial }_{P}f\left( x\right) \) is nonempty). 7.3. Theorem. Let \( U \subset X \) be open and convex, and let \( f \in \mathcal{F}\left( U\right) \) . Then \( f \) is Lipschitz on \( U \) of rank \( K \geq 0 \) iff \[ \parallel \zeta \parallel \leq K\forall \zeta \in {\partial }_{P}f\left( x\right) ,\forall x \in U. \] (1) Proof. If \( f \) has the stated Lipschitz property, then (1) follows, as already noted in Exercise 2.13. So let us now posit (1) and prove
1012_(GTM178)Nonsmooth Analysis and Control Theory
18
in \operatorname{int}\operatorname{co}\left\{ {p}_{i}\right\} \subset U \) .) The class of Lipschitz functions constitutes an extremely interesting one in its own right. We now show in the general setting that the Lipschitz property can be characterized in proximal terms. It is interesting to compare the nature of this proof to the very different proofs in the smooth case, which use integration or the Mean Value Theorem. Here, optimization is the principal tool, and the criterion need only hold at certain points (where \( {\partial }_{P}f\left( x\right) \) is nonempty). 7.3. Theorem. Let \( U \subset X \) be open and convex, and let \( f \in \mathcal{F}\left( U\right) \) . Then \( f \) is Lipschitz on \( U \) of rank \( K \geq 0 \) iff \[ \parallel \zeta \parallel \leq K\forall \zeta \in {\partial }_{P}f\left( x\right) ,\forall x \in U. \] (1) Proof. If \( f \) has the stated Lipschitz property, then (1) follows, as already noted in Exercise 2.13. So let us now posit (1) and prove that the Lipschitz property holds. We claim that it suffices to prove the following local property: for each \( {x}_{0} \in U \) there is an open ball centered at \( {x}_{0} \) on which \( f \) is Lipschitz of rank \( K \) . For suppose this local condition holds, and let \( x \) and \( y \) be any two points in \( U \) . Each point \( z \) of the line segment \( \left\lbrack {x, y}\right\rbrack \) admits a ball \( B\left( {z;r}\right) \) on which the Lipschitz condition holds. Since \( \left\lbrack {x, y}\right\rbrack \) is compact, a finite number of balls \( B\left( {{z}_{j};{r}_{j}}\right) \) covers \( \left\lbrack {x, y}\right\rbrack \) . This allows us to find points \( x + {t}_{i}\left( {y - x}\right) \) in \( \left\lbrack {x, y}\right\rbrack \left( {i = 0,1\ldots, N}\right) \), with \( 0 = {t}_{0} < {t}_{1} < \cdots < {t}_{N} = 1 \) , such that successive points always lie in one of the sets \( B\left( {{z}_{j};{r}_{j}}\right) \) . Then \[ f\left( y\right) - f\left( x\right) = \mathop{\sum }\limits_{{i = 0}}^{{N - 1}}\left\lbrack {f\left( {x + {t}_{i + 1}\left( {y - x}\right) }\right) - f\left( {x + {t}_{i}\left( {y - x}\right) }\right) }\right\rbrack \] \[ \leq \mathop{\sum }\limits_{{i = 0}}^{{N - 1}}K\left( {{t}_{i + 1} - {t}_{i}}\right) \parallel y - x\parallel = K\parallel y - x\parallel \] which confirms the global Lipschitz condition on \( U \) . We turn now to verifying the local property. The first step will consist of showing that each point \( {x}_{0} \in \operatorname{cl}\left( {{\operatorname{dom}}_{U}f}\right) \cap U \) admits a neighborhood in which \( f \) is finite and Lipschitz of rank \( K \) . Here, \( {\operatorname{dom}}_{U}f \) means the set of those points in \( U \) at which \( f \) is finite, a set which is nonempty by definition of \( \mathcal{F}\left( U\right) \) . Let \( \varepsilon > 0 \) be such that \( f \) is bounded below on \( \bar{B}\left( {{x}_{0};{4\varepsilon }}\right) \subset U \) ; pick any \( {K}^{\prime } > K \) . We denote by \( \varphi \) a function mapping the interval \( \lbrack 0,{3\varepsilon }) \) in \( \mathbb{R} \) to \( \lbrack 0,\infty ) \) and having the following properties: \( \varphi \left( \cdot \right) \) is strictly increasing, \( \varphi \left( t\right) = {K}^{\prime }t \) for \( 0 \leq t \leq {2\varepsilon },{\varphi }^{\prime }\left( t\right) \geq {K}^{\prime } \) for \( t \geq {2\varepsilon },\varphi \left( \cdot \right) \) is \( {C}^{2} \) on \( \left( {0,{3\varepsilon }}\right) \) , \( \varphi \left( t\right) \rightarrow \infty \) as \( t \rightarrow {3\varepsilon } \) . (A sketch will show the reader what is involved.) Now fix any two points \( y, z \in B\left( {{x}_{0};\varepsilon }\right) \) . The function \[ x \mapsto f\left( {y + x}\right) + \varphi \left( {\parallel x\parallel }\right) \] is not identically \( + \infty \) on \( B\left( {0;{3\varepsilon }}\right) \), since \( {x}_{0} \in \operatorname{cl}\left( {{\operatorname{dom}}_{U}f}\right) \) . It is lower semicontinuous and bounded below on \( \bar{B}\left( {0;{3\varepsilon }}\right) \), and is equal to \( + \infty \) on the boundary of that closed ball. We invoke the minimization principle embodied in Theorem 4.1 to deduce that for some \( \beta \in X \) with \( \parallel \beta \parallel < {K}^{\prime } - K \), the function \( g \) defined as \[ g\left( x\right) \mathrel{\text{:=}} f\left( {y + x}\right) + \varphi \left( {\parallel x\parallel }\right) - \langle \beta, x\rangle \] attains a minimum over \( \bar{B}\left( {0;{3\varepsilon }}\right) \) at a point \( u \) . Necessarily we have \( \parallel u\parallel < {3\varepsilon } \) , so that \( 0 \in {\partial }_{P}g\left( u\right) \) . If \( u \neq 0 \), then the function \( x \mapsto \varphi \left( {\parallel x\parallel }\right) \) is \( {C}^{2} \) in a neighborhood of \( u \), and so Proposition 2.11 applies to give \[ \beta - {\varphi }^{\prime }\left( {\parallel u\parallel }\right) u/\parallel u\parallel \in {\partial }_{P}f\left( {y + u}\right) . \] But \[ \begin{Vmatrix}{\beta - {\varphi }^{\prime }\left( {\parallel u\parallel }\right) u/\parallel u\parallel }\end{Vmatrix} \geq {\varphi }^{\prime }\left( {\parallel u\parallel }\right) - \parallel \beta \parallel \] \[ > {K}^{\prime } - \left( {{K}^{\prime } - K}\right) = K \] and \( y + u \) is a point in \( B\left( {{x}_{0};{4\varepsilon }}\right) \subset U \) . This contradicts the given bound on \( {\partial }_{P}f \) and implies that \( u = 0 \) necessarily. Since \( u = 0 \) minimizes \( g \) over \( B\left( {0;{3\varepsilon }}\right) \), we deduce that \( f\left( y\right) < \infty \), and \[ f\left( y\right) = g\left( 0\right) \leq g\left( {z - y}\right) \] \[ = f\left( z\right) + \varphi \left( {\parallel z - y\parallel }\right) - \langle \beta, z - y\rangle \] \[ \leq f\left( z\right) + {K}^{\prime }\begin{Vmatrix}{z - y}\end{Vmatrix} + \left( {{K}^{\prime } - K}\right) \parallel z - y\parallel \] where we have used the fact that \( \parallel z - y\parallel < {2\varepsilon } \) and the fact that \( \varphi \left( t\right) = {K}^{\prime }t \) for \( t \in \left\lbrack {0,{2\varepsilon }}\right\rbrack \) . Since \( y \) and \( z \) are arbitrary points in \( B\left( {{x}_{0};\varepsilon }\right) \) and \( {K}^{\prime } > K \) is arbitrary too, we have shown that \( f \) is Lipschitz of rank \( K \) on \( B\left( {{x}_{0};\varepsilon }\right) \) . There remains to deal with the restriction that \( {x}_{0} \) belongs to \( \operatorname{cl}\left( {{\operatorname{dom}}_{U}f}\right) \) . But the argument just given evidently implies that any such \( {x}_{0} \) belongs to \( \operatorname{int}\left( {{\operatorname{dom}}_{U}f}\right) \), so that \( \operatorname{cl}\left( {{\operatorname{dom}}_{U}f}\right) \) is seen to be open (as well as closed) relative to \( U \) ; consequently \( \operatorname{cl}\left( {{\operatorname{dom}}_{U}f}\right) \cap U = {\operatorname{dom}}_{U}f = U \) . Therefore the local property holds at every point of \( U \) and the proof is complete. 7.4. Corollary. Let \( U \subset X \) be open and convex, and let \( f \in \mathcal{F}\left( U\right) \) . Then \( f \) is constant on \( U \) iff \[ {\partial }_{P}f\left( x\right) \subset \{ 0\} \forall x \in U. \] 7.5. Exercise. (a) Show that Corollary 7.4 is false if the hypothesis that \( U \) be convex is deleted. (b) Show that if \( U \) is open (not necessarily convex), then \( f \in \mathcal{F}\left( U\right) \) is locally Lipschitz on \( U \) iff \( {\partial }_{P}f \) is locally bounded on \( U \) . (c) Prove that a function \( f \) is locally Lipschitz on a compact set \( S \) iff \( f \) is globally Lipschitz on \( S \) . ## 8 The Sum Rule Suppose that the sum of two functions \( {f}_{1} + {f}_{2} \) is minimized at a point \( {x}_{0} \) : \[ {f}_{1}\left( x\right) + {f}_{2}\left( x\right) \geq {f}_{1}\left( {x}_{0}\right) + {f}_{2}\left( {x}_{0}\right) \forall x \in X. \] Then of course we have \( 0 \in {\partial }_{P}\left( {{f}_{1} + {f}_{2}}\right) \left( {x}_{0}\right) \) . Do we also have the "separated conclusion" \( 0 \in {\partial }_{P}{f}_{1}\left( {x}_{0}\right) + {\partial }_{P}{f}_{2}\left( {x}_{0}\right) \) ? That the answer in general is "no" is easy to see: take, for example, \( {f}_{1}\left( x\right) \mathrel{\text{:=}} \left| x\right| \) and \( {f}_{2}\left( x\right) \mathrel{\text{:=}} - \left| x\right| \) on \( X = \mathbb{R} \) , with \( {x}_{0} = 0 \) . Then \( {\partial }_{P}{f}_{2}\left( {x}_{0}\right) = \varnothing \), so that the separated conclusion cannot hold. More generally, an exact Sum Rule of the form \( {\partial }_{P}\left( {{f}_{1} + {f}_{2}}\right) \left( x\right) = \) \( {\partial }_{P}{f}_{1}\left( x\right) + {\partial }_{P}{f}_{2}\left( x\right) \) is not available. On the other hand, if \( {f}_{1} \) and \( {f}_{2} \) are functions of two different unrelated arguments, that is, if we have \( {f}_{1}\left( x\right) + {f}_{2}\left( y\right) \), then at a minimizing pair \( \left( {{x}_{0},{y}_{0}}\right) \) of \( {f}_{1}\left( x\right) + {f}_{2}\left( y\right) \) we evidently have the separate conclusions \( 0 \in {\partial }_{P}{f}_{1}\left( {x}_{0}\right) \) and \( 0 \in {\partial }_{P}{f}_{2}\left( {y}_{0}\right) \) . There is a technique, sometimes called "decoupling," to approximate the first (coupled) situation above by the second (uncoupled) one. To apply this technique, we penalize the pairs \( \left( {x, y}\right) \) with \( x \neq y \) by adding a large positive multiple of \( \parallel x - y{\parallel }^{2} \) ; that is, we consider minimizing \[ {f}_{1}\left( x\right) + {f}_{2}\left( y\right) + r\parallel x - y{\parallel }^{2} \] over the pairs \( \left( {x, y}\right) \) in \( X \times X \) . To make the decoupling technique work, we need to know that the minimum approaches that of the coupled case as \( r \rightarrow \infty \) . We will give a precise result of this type, whose relevance to the Sum Rule will be appa
1012_(GTM178)Nonsmooth Analysis and Control Theory
19
at a minimizing pair \( \left( {{x}_{0},{y}_{0}}\right) \) of \( {f}_{1}\left( x\right) + {f}_{2}\left( y\right) \) we evidently have the separate conclusions \( 0 \in {\partial }_{P}{f}_{1}\left( {x}_{0}\right) \) and \( 0 \in {\partial }_{P}{f}_{2}\left( {y}_{0}\right) \) . There is a technique, sometimes called "decoupling," to approximate the first (coupled) situation above by the second (uncoupled) one. To apply this technique, we penalize the pairs \( \left( {x, y}\right) \) with \( x \neq y \) by adding a large positive multiple of \( \parallel x - y{\parallel }^{2} \) ; that is, we consider minimizing \[ {f}_{1}\left( x\right) + {f}_{2}\left( y\right) + r\parallel x - y{\parallel }^{2} \] over the pairs \( \left( {x, y}\right) \) in \( X \times X \) . To make the decoupling technique work, we need to know that the minimum approaches that of the coupled case as \( r \rightarrow \infty \) . We will give a precise result of this type, whose relevance to the Sum Rule will be apparent shortly. Supplementary hypotheses are required, and we identify two alternatives, one of which involves weak lower semicontinuity. We say that a function \( f \) is weakly lower semicontinuous at \( x \) provided that \[ \mathop{\liminf }\limits_{{i \rightarrow \infty }}f\left( {x}_{i}\right) \geq f\left( x\right) \] whenever the sequence \( {x}_{i} \) converges weakly to \( x \) . (Note the sequential nature of this condition.) 8.1. Exercise. (a) Let \( f \in \mathcal{F} \) be convex. Prove that \( f \) is weakly lower semicontinuous at every point \( x \) . (Hint. Recall that a convex subset of a Hilbert space is closed iff it is weakly closed iff it is sequentially weakly closed.) (b) Let \( S \) be a weakly closed subset of \( X \) . Then its indicator function \( {I}_{S} \) is weakly lower semicontinuous. 8.2. Proposition. Let \( {f}_{1} \) and \( {f}_{2} \) belong to \( \mathcal{F}\left( X\right) \), and let \( C \) be a closed convex bounded subset of \( X \) with \( C \cap \operatorname{dom}{f}_{1} \cap \operatorname{dom}{f}_{2} \neq \varnothing \) . Suppose that either: (i) \( {f}_{1} \) and \( {f}_{2} \) are weakly lower semicontinuous on \( C \) (automatically the case if \( X \) is finite dimensional); or (ii) one of the functions is Lipschitz on \( C \) and the other is bounded below on \( C \) . Then, for any positive sequence \( \left\{ {r}_{n}\right\} \) with \( \mathop{\lim }\limits_{{n \rightarrow \infty }}{r}_{n} = + \infty \), we have \[ \mathop{\lim }\limits_{{n \rightarrow \infty }}\mathop{\inf }\limits_{{x, y \in C}}\left\{ {{f}_{1}\left( x\right) + {f}_{2}\left( y\right) + {r}_{n}\parallel x - y{\parallel }^{2}}\right\} = \mathop{\inf }\limits_{{x \in C}}\left\{ {{f}_{1}\left( x\right) + {f}_{2}\left( x\right) }\right\} \] (1) Proof. The left side of (1) is no greater than the right, so only the opposite inequality need be proved. We address first the case in which hypothesis (i) holds. In that case, since \( C \) is weakly compact and since \( \left( {x, y}\right) \mapsto \parallel x - y{\parallel }^{2} \) is convex on \( X \times X \) (and hence, weakly lower semicontinuous), the function \( \left( {x, y}\right) \rightarrow {f}_{1}\left( x\right) + {f}_{2}\left( y\right) + {r}_{n}\parallel x - y{\parallel }^{2} \) admits a minimum over \( C \times C \), at a point we will label \( \left( {{x}_{n},{y}_{n}}\right) \) . Let us extract a weakly convergent subsequence from \( \left\{ \left( {{x}_{n},{y}_{n}}\right) \right\} \), without relabeling. If \( {x}_{0} \) is any point in \( C \cap \operatorname{dom}{f}_{1} \cap \operatorname{dom}{f}_{2} \), then we have \[ {f}_{1}\left( {x}_{n}\right) + {f}_{2}\left( {y}_{n}\right) + {r}_{n}{\begin{Vmatrix}{x}_{n} - {y}_{n}\end{Vmatrix}}^{2} \leq {f}_{1}\left( {x}_{0}\right) + {f}_{2}\left( {x}_{0}\right) \] which yields an a priori bound: \( {r}_{n}{\begin{Vmatrix}{x}_{n} - {y}_{n}\end{Vmatrix}}^{2} \leq m \) . It follows that \( \parallel {x}_{n} - \) \( {y}_{n}\parallel \rightarrow 0 \) as \( n \rightarrow \infty \), from which we deduce that the weakly convergent sequences \( \left\{ {x}_{n}\right\} \) and \( \left\{ {y}_{n}\right\} \) have the same weak limit \( \bar{x} \in C \) . Invoking weak lower semicontinuity now leads to \[ {f}_{1}\left( \bar{x}\right) + {f}_{2}\left( \bar{x}\right) \leq \mathop{\liminf }\limits_{{n \rightarrow \infty }}\left\{ {{f}_{1}\left( {x}_{n}\right) + {f}_{2}\left( {y}_{n}\right) + {r}_{n}{\begin{Vmatrix}{x}_{n} - {y}_{n}\end{Vmatrix}}^{2}}\right\} \] which implies that the right side of (1) does not exceed the left. We turn now to the case of hypothesis (ii), letting \( K \) be a Lipschitz constant for one of the functions on \( C \), let us say \( {f}_{1} \) . For each \( n \), let \( \left( {{x}_{n},{y}_{n}}\right) \in C \times C \) satisfy \[ {f}_{1}\left( {x}_{n}\right) + {f}_{2}\left( {y}_{n}\right) + {r}_{n}{\begin{Vmatrix}{x}_{n} - {y}_{n}\end{Vmatrix}}^{2} \leq \mathop{\inf }\limits_{{x, y \in C}}\left\{ {{f}_{1}\left( x\right) + {f}_{2}\left( y\right) + {r}_{n}\parallel x - y{\parallel }^{2}}\right\} + \frac{1}{n}. \] It follows essentially as in the first case above that \( \begin{Vmatrix}{{x}_{n} - {y}_{n}}\end{Vmatrix} \rightarrow 0 \) as \( n \rightarrow \infty \) , since \( {f}_{1} \) and \( {f}_{2} \) are bounded below on \( C \) . We argue now as follows: \[ \mathop{\inf }\limits_{{x, y \in C}}\left\{ {{f}_{1}\left( x\right) + {f}_{2}\left( y\right) + {r}_{n}\parallel x - y{\parallel }^{2}}\right\} \] \[ \geq {f}_{1}\left( {x}_{n}\right) + {f}_{2}\left( {y}_{n}\right) + {r}_{n}{\begin{Vmatrix}{x}_{n} - {y}_{n}\end{Vmatrix}}^{2} - \frac{1}{n} \] \[ \geq {f}_{1}\left( {y}_{n}\right) - K\begin{Vmatrix}{{y}_{n} - {x}_{n}}\end{Vmatrix} + {f}_{2}\left( {y}_{n}\right) + {r}_{n}{\begin{Vmatrix}{x}_{n} - {y}_{n}\end{Vmatrix}}^{2} - \frac{1}{n} \] \[ \geq \mathop{\inf }\limits_{{x \in C}}\left\{ {{f}_{1}\left( x\right) + {f}_{2}\left( x\right) }\right\} - K\begin{Vmatrix}{{y}_{n} - {x}_{n}}\end{Vmatrix} - \frac{1}{n}. \] Passing to the limit as \( n \rightarrow \infty \) yields the result. We are now ready to prove a result known as the "fuzzy Sum Rule." 8.3. Theorem. Let \( {x}_{0} \in \operatorname{dom}{f}_{1} \cap \operatorname{dom}{f}_{2} \), and let \( \zeta \) belong to \( {\partial }_{P}\left( {{f}_{1} + }\right. \) \( \left. {f}_{2}\right) \left( {x}_{0}\right) \) . Suppose that either: (i) \( {f}_{1} \) and \( {f}_{2} \) are weakly lower semicontinuous (automatically the case if \( X \) is finite dimensional); or (ii) one of the functions is Lipschitz near \( {x}_{0} \) . Then, for any \( \varepsilon > 0 \), there exist (for \( i = 1,2 \) ) points \( {x}_{i} \in B\left( {{x}_{0};\varepsilon }\right) \) with \( \left| {{f}_{i}\left( {x}_{0}\right) - {f}_{i}\left( {x}_{i}\right) }\right| < \varepsilon \) such that \[ \zeta \in {\partial }_{P}{f}_{1}\left( {x}_{1}\right) + {\partial }_{P}{f}_{2}\left( {x}_{2}\right) + {\varepsilon B}. \] Proof. Since \( \zeta \) belongs to \( {\partial }_{P}\left( {{f}_{1} + {f}_{2}}\right) \left( {x}_{0}\right) \), there exists a closed ball \( C \mathrel{\text{:=}} \) \( \bar{B}\left( {{x}_{0};\eta }\right) \) of positive radius \( \eta \) centered at \( {x}_{0} \) and \( \sigma \geq 0 \) such that the function \[ x \mapsto {f}_{1}\left( x\right) + {f}_{2}\left( x\right) + \sigma {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2} - \left\langle {\zeta, x - {x}_{0}}\right\rangle \] attains a minimum over \( \bar{B}\left( {{x}_{0};\eta }\right) \) at \( x = {x}_{0} \) . We can take \( \eta \) small enough so that both \( {f}_{1} \) and \( {f}_{2} \) are bounded below on \( C \) and so that in case (ii), one of the functions is Lipschitz on that set. Now consider the problem of minimizing over \( X \times X \) the function \[ {\varphi }_{n}\left( {x, y}\right) : = {f}_{1}\left( x\right) + {f}_{2}\left( y\right) + \sigma {\begin{Vmatrix}y - {x}_{0}\end{Vmatrix}}^{2} - \left\langle {\zeta, x - {x}_{0}}\right\rangle + {I}_{C}\left( x\right) + {I}_{C}\left( y\right) + {r}_{n}{\begin{Vmatrix}x - y\end{Vmatrix}}^{2}, \] where \( {r}_{n} \) is a positive sequence tending to \( + \infty \) as \( n \rightarrow \infty \) . Applying Proposition 8.2 (with the role of \( {f}_{2}\left( y\right) \) played here by \[ \left. {{f}_{2}\left( y\right) + \sigma {\begin{Vmatrix}y - {x}_{0}\end{Vmatrix}}^{2} - \left\langle {\zeta, x - {x}_{0}}\right\rangle }\right) \] yields the conclusion that the nonnegative quantity \[ {q}_{n} \mathrel{\text{:=}} {f}_{1}\left( {x}_{0}\right) + {f}_{2}\left( {x}_{0}\right) - \mathop{\inf }\limits_{{X \times X}}{\varphi }_{n} \] tends to zero as \( n \rightarrow \infty \) . Put another way, we conclude that for \( n \) large, the point \( \left( {{x}_{0},{x}_{0}}\right) \) "almost minimizes" \( {\varphi }_{n} \) . We will invoke the Minimization Principle Theorem 4.2 with data \( f \mathrel{\text{:=}} {\varphi }_{n} \) , \( \varepsilon \mathrel{\text{:=}} {q}_{n} + 1/n,\lambda \mathrel{\text{:=}} \sqrt{\varepsilon } \) . We derive the existence of points \( \left( {{x}_{1}^{n},{x}_{2}^{n}}\right) \left( { = y}\right) \) and \( \left( {{z}_{1}^{n},{z}_{2}^{n}}\right) \left( { = z}\right) \) in \( X \times X \) such that \[ \begin{Vmatrix}{\left( {{x}_{1}^{n},{x}_{2}^{n}}\right) - \left( {{z}_{1}^{n},{z}_{2}^{n}}\right) }\end{Vmatrix} < {\left( {q}_{n} + \frac{1}{n}\right) }^{1/2},\;\begin{Vmatrix}{\left( {{z}_{1}^{n},{z}_{2}^{n}}\right) - \left( {{x}_{0},{x}_{0}}\right) }\end{Vmatrix} < {\left( {q}_{n} + \frac{1}{n}\right) }^{1/2}, \] and such that the function \[ {\varphi }_{n}\left( {x, y}\right) + {\begin{Vmatrix}x - {z}_{1}^{n}\end{Vmatrix}}^{2} + {\begin{Vmatrix}y - {z}_{2}^{n}\end{Vmatrix}}^{2} \] is minimized over \( X \times X \) at \( \left( {x, y}\right) = \left( {{x}_{1}^{n},{x}_{2}^{n}}\right) \) . For \( n \) sufficiently large (since \( \left. {{q}_{n} \rightarrow 0}\right) \), both \( {x}_{1}^{n} \) and \( {x}_{2
1012_(GTM178)Nonsmooth Analysis and Control Theory
20
{:=}} {q}_{n} + 1/n,\lambda \mathrel{\text{:=}} \sqrt{\varepsilon } \) . We derive the existence of points \( \left( {{x}_{1}^{n},{x}_{2}^{n}}\right) \left( { = y}\right) \) and \( \left( {{z}_{1}^{n},{z}_{2}^{n}}\right) \left( { = z}\right) \) in \( X \times X \) such that \[ \begin{Vmatrix}{\left( {{x}_{1}^{n},{x}_{2}^{n}}\right) - \left( {{z}_{1}^{n},{z}_{2}^{n}}\right) }\end{Vmatrix} < {\left( {q}_{n} + \frac{1}{n}\right) }^{1/2},\;\begin{Vmatrix}{\left( {{z}_{1}^{n},{z}_{2}^{n}}\right) - \left( {{x}_{0},{x}_{0}}\right) }\end{Vmatrix} < {\left( {q}_{n} + \frac{1}{n}\right) }^{1/2}, \] and such that the function \[ {\varphi }_{n}\left( {x, y}\right) + {\begin{Vmatrix}x - {z}_{1}^{n}\end{Vmatrix}}^{2} + {\begin{Vmatrix}y - {z}_{2}^{n}\end{Vmatrix}}^{2} \] is minimized over \( X \times X \) at \( \left( {x, y}\right) = \left( {{x}_{1}^{n},{x}_{2}^{n}}\right) \) . For \( n \) sufficiently large (since \( \left. {{q}_{n} \rightarrow 0}\right) \), both \( {x}_{1}^{n} \) and \( {x}_{2}^{n} \) must lie in the interior of \( C = \bar{B}\left( {{x}_{0};\eta }\right) \), so that the fact of this minimization implies the separate necessary conditions \[ - 2{r}_{n}\left( {{x}_{1}^{n} - {x}_{2}^{n}}\right) - 2\left( {{x}_{1}^{n} - {z}_{1}^{n}}\right) \in {\partial }_{P}{f}_{1}\left( {x}_{1}^{n}\right) \text{ (minimum in }x\text{ ),} \] \[ 2{r}_{n}\left( {{x}_{1}^{n} - {x}_{2}^{n}}\right) - {2\sigma }\left( {{x}_{2}^{n} - {x}_{0}}\right) - 2\left( {{x}_{2}^{n} - {z}_{2}^{n}}\right) \in {\partial }_{P}{f}_{2}\left( {x}_{2}^{n}\right) \text{ (minimum in }y\text{ ). } \] We derive from this \[ \zeta \in {\partial }_{P}{f}_{1}\left( {x}_{1}^{n}\right) + {\partial }_{P}{f}_{2}\left( {x}_{2}^{n}\right) + {\gamma }_{n}\bar{B} \] where \[ {\gamma }_{n} \mathrel{\text{:=}} 2\left( {\begin{Vmatrix}{{x}_{1}^{n} - {z}_{1}^{n}}\end{Vmatrix} + \begin{Vmatrix}{{x}_{2}^{n} - {z}_{2}^{n}}\end{Vmatrix} + \sigma \begin{Vmatrix}{{x}_{2}^{n} - {x}_{0}}\end{Vmatrix}}\right) . \] For all \( n \) large enough, the points \( {x}_{1}^{n} \) and \( {x}_{2}^{n} \) will lie within the pregiven \( \varepsilon \) of \( {x}_{0} \), and \( {\gamma }_{n} \) will be less than \( \varepsilon \) too. So all that remains to check is the closeness of each of the function values \( {f}_{i}\left( {x}_{i}^{n}\right) \) to \( {f}_{i}\left( {x}_{0}\right) \left( {i = 1,2}\right) \), for \( n \) sufficiently large. This will follow from the further assertion of Theorem 4.2 that we have \[ {\varphi }_{n}\left( {{x}_{1}^{n},{x}_{2}^{n}}\right) \leq {\varphi }_{n}\left( {{x}_{0},{x}_{0}}\right) \] which implies that for all \( n \) large the following holds: \[ {f}_{1}\left( {x}_{1}^{n}\right) + {f}_{2}\left( {x}_{2}^{n}\right) \leq {f}_{1}\left( {x}_{0}\right) + {f}_{2}\left( {x}_{0}\right) + \parallel \zeta \parallel \begin{Vmatrix}{{x}_{1}^{n} - {x}_{0}}\end{Vmatrix}. \] We derive from this \[ \mathop{\liminf }\limits_{{n \rightarrow \infty }}\left\{ {{f}_{1}\left( {x}_{1}^{n}\right) + {f}_{2}\left( {x}_{2}^{n}\right) }\right\} \leq {f}_{1}\left( {x}_{0}\right) + {f}_{2}\left( {x}_{0}\right) . \] In view of the lower semicontinuity of \( {f}_{1} \) and \( {f}_{2} \), this yields the fact that for \( i = 1,2 \) the sequence \( {f}_{i}\left( {x}_{i}^{n}\right) \) converges to \( {f}_{i}\left( {x}_{0}\right) \) as \( n \rightarrow \infty \) (why?), completing the proof of the theorem. 8.4. Exercise. (a) Let \( {C}_{1} \) and \( {C}_{2} \) be weakly closed subsets of \( X \), and let \( \zeta \in \) \( {N}_{{C}_{1} \cap {C}_{2}}^{P}\left( x\right) \) . Then for any \( \varepsilon > 0 \) there exist \[ {x}_{1} \in {C}_{1},\;{x}_{2} \in {C}_{2},\;{\zeta }_{1} \in {N}_{{C}_{1}}^{P}\left( {x}_{1}\right) ,\;{\zeta }_{2} \in {N}_{{C}_{2}}^{P}\left( {x}_{2}\right) , \] such that \[ \begin{Vmatrix}{{x}_{1} - x}\end{Vmatrix} + \begin{Vmatrix}{{x}_{2} - x}\end{Vmatrix} < \varepsilon ,\;\begin{Vmatrix}{\zeta - {\zeta }_{1} - {\zeta }_{2}}\end{Vmatrix} < \varepsilon . \] (b) Suppose that \( x \) minimizes the locally Lipschitz function \( f \) over the set \( C \) . Then for any \( \varepsilon > 0 \) there exist \( {x}_{1} \) and \( {x}_{2} \) in the \( \varepsilon \) -neighborhood of \( x \) such that \[ 0 \in {\partial }_{P}f\left( {x}_{1}\right) + {N}_{C}^{P}\left( {x}_{2}\right) + {\varepsilon B}. \] ## 9 The Chain Rule Suppose now that the function \( f \) is given as the composition \( g \circ F \) of two functions \( g : Y \rightarrow ( - \infty ,\infty \rbrack \) and \( F : X \rightarrow Y \), where \( Y \) is another Hilbert space. Note that if \( g \) belongs to \( \mathcal{F}\left( Y\right) \) and \( F \) is continuous (in particular, if \( F \) is locally Lipschitz, as we will assume), then it follows readily that \( f = g \circ F \) belongs to \( \mathcal{F}\left( X\right) \) . In the classical smooth setting the Chain Rule asserts \( {f}^{\prime }\left( x\right) = {g}^{\prime }\left( {F\left( x\right) }\right) \circ {F}^{\prime }\left( x\right) \) . In extending this to the nonsmooth setting, one of the apparent difficulties is that we have not defined a replacement for the derivative \( {F}^{\prime }\left( x\right) \) of a nondifferentiable vector-valued mapping such as \( F \) . We will get around this by a device which, in the smooth case, amounts to viewing \( {f}^{\prime }\left( x\right) \) as the derivative of the scalar-valued function \( u \rightarrow {g}^{\prime }\left( {F\left( x\right) }\right) \circ F\left( u\right) \) . Also notable about our approach is that it reveals the composition of functions to be reducible to the case of the sum of two functions. 9.1. Theorem. Let \( g \in \mathcal{F}\left( Y\right) \), and let \( F : X \rightarrow Y \) be locally Lipschitz. Set \( f\left( x\right) \mathrel{\text{:=}} g\left( {F\left( x\right) }\right) \), and let \( \zeta \in {\partial }_{P}f\left( {x}_{0}\right) \) . We assume that either: (i) \( g \) is weakly lower semicontinuous and \( F \) is linear; or (ii) \( g \) is Lipschitz near \( F\left( {x}_{0}\right) \) . Then for all \( \varepsilon > 0 \) there exist \( \widetilde{x} \) in \( {x}_{0} + \varepsilon {B}_{X},\widetilde{y} \) in \( F\left( {x}_{0}\right) + \varepsilon {B}_{Y} \), and \( \gamma \in {\partial }_{P}g\left( \widetilde{y}\right) \) such that \( \begin{Vmatrix}{F\left( \widetilde{x}\right) - F\left( {x}_{0}\right) }\end{Vmatrix} < \varepsilon \) and \[ \zeta \in {\partial }_{P}\{ \langle \gamma, F\left( \cdot \right) \rangle \} \left( \widetilde{x}\right) + \varepsilon {B}_{X} \] Proof. Since \( \zeta \) belongs to \( {\partial }_{P}f\left( {x}_{0}\right) \), there exist \( \sigma ,\eta > 0 \) such that \[ f\left( x\right) - \langle \zeta, x\rangle + \sigma {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2} \geq f\left( {x}_{0}\right) - \left\langle {\zeta ,{x}_{0}}\right\rangle \] for all \( x \in B\left( {{x}_{0};\eta }\right) \) . Let \( S \) denote the graph of \( F \) in the space \( X \times Y \) . Then another way of writing the preceding proximal subgradient inequality is the following: \[ g\left( y\right) - \langle \zeta, x\rangle + {I}_{S}\left( {x, y}\right) + \sigma {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2} \geq g\left( {F\left( {x}_{0}\right) }\right) - \left\langle {\zeta ,{x}_{0}}\right\rangle , \] which implies that the left side, as a function of \( \left( {x, y}\right) \), attains a local minimum at \( \left( {x, y}\right) = \left( {{x}_{0}, F\left( {x}_{0}\right) }\right) \) . We deduce \[ 0 \in {\partial }_{P}\left\{ {g\left( y\right) -\langle \zeta, x\rangle + {I}_{S}\left( {x, y}\right) + \sigma {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2}}\right\} \left( {{x}_{0}, F\left( {x}_{0}\right) }\right) . \] In each of cases (i) and (ii) of the theorem, we are justified in applying the fuzzy Sum Rule, Theorem 8.3. We obtain, for any \( \delta > 0 \), the existence of points \( \left( {{x}_{1},{y}_{1}}\right) \) and \( \left( {{x}_{2},{y}_{2}}\right) \), both \( \delta \) -near \( \left( {{x}_{0}, F\left( {x}_{0}\right) }\right) \), such that, for some \( \gamma \in {\partial }_{P}g\left( {y}_{1}\right) \) we have \[ \left( {\zeta - {2\sigma }\left( {{x}_{1} - {x}_{0}}\right) , - \gamma }\right) \in {\partial }_{P}{I}_{S}\left( {{x}_{2},{y}_{2}}\right) + \delta {B}_{X \times Y} \] But \( {\partial }_{P}{I}_{S}\left( {{x}_{2},{y}_{2}}\right) = {N}_{S}^{P}\left( {{x}_{2},{y}_{2}}\right) \) and it follows that \( {y}_{2} = F\left( {x}_{2}\right) \) and, from Exercise 9.2 below, that for some \( u, v \) with \( \parallel u\parallel < \delta ,\parallel v\parallel < \delta \), we have \[ \zeta - {2\sigma }\left( {{x}_{1} - {x}_{0}}\right) + u \in {\partial }_{P}\{ \langle \gamma + v, F\left( \cdot \right) \rangle \} \left( {x}_{2}\right) . \] Now let us apply the fuzzy Sum Rule again to the functions \( \langle \gamma, F\left( \cdot \right) \rangle \) and \( \langle v, F\left( \cdot \right) \rangle \), which are locally Lipschitz. We deduce the existence of points \( {x}_{3} \) and \( {x}_{4} \) in the \( \delta \) -neighborhood of \( {x}_{2} \) such that \[ \zeta - {2\sigma }\left( {{x}_{1} - {x}_{0}}\right) + u \in {\partial }_{P}\{ \langle \gamma, F\left( \cdot \right) \rangle \} \left( {x}_{3}\right) + {\partial }_{P}\{ \langle v, F\left( \cdot \right) \rangle \} \left( {x}_{4}\right) + \delta {B}_{X} \] If \( K \) is a local Lipschitz constant for \( F \), then the function \( \langle v, F\left( \cdot \right) \rangle \) is locally Lipschitz of rank \( K\parallel v\parallel < {K\delta } \), whence any element of \( {\partial }_{P}\{ \langle v, F\left( \cdot \right) \rangle \} \left( {x}_{4}\right) \) is of norm less than \( {K\delta } \) . We have now arrived at \[ \zeta - {2\sigma }\left( {{x}_{1} - {x}_{0}}\right) + u \in {\partial }_{P}\{ \langle \gamma, F\left( \cdot \right) \rangle \} \left( {x}_{3}\right) + \delta \left( {K + 1}\right) {B}_{X}. \] It suffices now to take \( \delta \) small enough to get the required result, wit
1012_(GTM178)Nonsmooth Analysis and Control Theory
21
ts \( {x}_{3} \) and \( {x}_{4} \) in the \( \delta \) -neighborhood of \( {x}_{2} \) such that \[ \zeta - {2\sigma }\left( {{x}_{1} - {x}_{0}}\right) + u \in {\partial }_{P}\{ \langle \gamma, F\left( \cdot \right) \rangle \} \left( {x}_{3}\right) + {\partial }_{P}\{ \langle v, F\left( \cdot \right) \rangle \} \left( {x}_{4}\right) + \delta {B}_{X} \] If \( K \) is a local Lipschitz constant for \( F \), then the function \( \langle v, F\left( \cdot \right) \rangle \) is locally Lipschitz of rank \( K\parallel v\parallel < {K\delta } \), whence any element of \( {\partial }_{P}\{ \langle v, F\left( \cdot \right) \rangle \} \left( {x}_{4}\right) \) is of norm less than \( {K\delta } \) . We have now arrived at \[ \zeta - {2\sigma }\left( {{x}_{1} - {x}_{0}}\right) + u \in {\partial }_{P}\{ \langle \gamma, F\left( \cdot \right) \rangle \} \left( {x}_{3}\right) + \delta \left( {K + 1}\right) {B}_{X}. \] It suffices now to take \( \delta \) small enough to get the required result, with \( \widetilde{y} = {y}_{1} \) and \( \widetilde{x} = {x}_{3} \) . The closeness of \( F\left( \widetilde{x}\right) = F\left( {x}_{3}\right) \) to \( F\left( {x}_{0}\right) \) follows from the continuity of \( F \) . Here is the result needed in the proof above: 9.2. Exercise. Let \( F : X \rightarrow Y \) be locally Lipschitz, and suppose that \[ \left( {\alpha , - \gamma }\right) \in {N}_{S}^{P}\left( {x, F\left( x\right) }\right) \] where \( S \) is the graph of \( F \) . Prove that \[ \alpha \in {\partial }_{P}\{ \langle \gamma, F\left( \cdot \right) \rangle \} \left( x\right) \] The two following exercises illustrate the use of the Chain Rule, and will be used later. 9.3. Exercise. Let \( g \in \mathcal{F}\left( Y\right) \) be weakly lower semicontinuous and let \( A : X \rightarrow Y \) be a continuous linear operator. Suppose that \( \zeta \) belongs to \( {\partial }_{P}\{ g\left( {Ax}\right) \} \left( {x}_{0}\right) \) . Prove the existence, for any \( \varepsilon > 0 \), of \( \widetilde{y} \) within \( \varepsilon \) of \( A{x}_{0} \) such that \[ \zeta \in {A}^{ * }{\partial }_{P}g\left( \widetilde{y}\right) + \varepsilon {B}_{X}\;\text{ and }\;\left| {g\left( \widetilde{y}\right) - g\left( {A{x}_{0}}\right) }\right| < \varepsilon . \] (Here \( {A}^{ * } : Y \rightarrow X \) denotes the adjoint of \( A : \left\langle {{A}^{ * }u, v}\right\rangle = \langle u,{Av}\rangle \) for all \( u \in Y, v \in X \) .) 9.4. Exercise. Let \( g \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) be locally Lipschitz, and define \( f \) as follows: \[ f : {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \rightarrow \mathbb{R},\;f\left( v\right) \mathrel{\text{:=}} g\left( {{\int }_{a}^{b}v\left( s\right) {ds}}\right) . \] Let \( \zeta \in {L}^{2} \) belong to \( {\partial }_{P}f\left( v\right) \) . Prove that for any \( \varepsilon > 0 \) there exist \( x \) and \( \theta \) with \( \theta \in {\partial }_{P}g\left( x\right) \) such that \[ \begin{Vmatrix}{x - {\int }_{a}^{b}v\left( s\right) {ds}}\end{Vmatrix} < \varepsilon ,\;\parallel \zeta \left( t\right) - \theta {\parallel }_{2} < \varepsilon . \] Armed with the Chain Rule above and the Sum Rule of the preceding section, we can easily derive other basic results of "proximal fuzzy calculus." But we turn instead to the issue of passing to the limit in results such as the above. ## 10 Limiting Calculus Consider once again the issue of a proximal Sum Rule. We have noted the easy fact \[ {\partial }_{P}{f}_{1}\left( x\right) + {\partial }_{P}{f}_{2}\left( x\right) \subseteq {\partial }_{P}\left( {{f}_{1} + {f}_{2}}\right) \left( x\right) \] (1) and we have proved an approximate form of the reverse inclusion: if \( \zeta \in \) \( {\partial }_{P}\left( {{f}_{1} + {f}_{2}}\right) \left( x\right) \), then for any \( \varepsilon > 0 \) we have \[ \zeta \in {\partial }_{P}{f}_{1}\left( {x}_{1}\right) + {\partial }_{P}{f}_{2}\left( {x}_{2}\right) + {\varepsilon B} \] (2) where \( {x}_{1} \) and \( {x}_{2} \) are \( \varepsilon \) -close to \( x \), and where \( {f}_{i}\left( {x}_{i}\right) \) and \( {f}_{i}\left( x\right) \) are close as well \( \left( {i = 1,2}\right) \) . It is natural to consider passing to the limit in (2) as \( \varepsilon \rightarrow 0 \) . In doing so, the outcome cannot necessarily be phrased in terms of \( {\partial }_{P}{f}_{1}\left( x\right) \) and \( {\partial }_{P}{f}_{2}\left( x\right) \) (which may be empty, for example). A construction that naturally recommends itself, and that we call the limiting subdifferential \( {\partial }_{L}f\left( x\right) \) (or \( L \) -subdifferential), is the following: \[ {\partial }_{L}f\left( x\right) \mathrel{\text{:=}} \left\{ {\mathrm{w} - \lim {\zeta }_{i} : {\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) ,{x}_{i}\overset{f}{ \rightarrow }x}\right\} . \] (3) That is, we consider the set of all vectors \( \zeta \) that can be expressed as the weak limit (which is what "w-lim" signifies) of some sequence \( \left\{ {\zeta }_{i}\right\} \), where \( {\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) \) for each \( i \), and where \( {x}_{i} \rightarrow x, f\left( {x}_{i}\right) \rightarrow f\left( x\right) \) . (Note: The notation \[ {x}_{i}\overset{f}{ \rightarrow }x \] encapsulates both of these convergences.) Let us return now to the Sum Rule, but in limiting terms: let \( \zeta \in {\partial }_{L}\left( {{f}_{1} + }\right. \) \( \left. {f}_{2}\right) \left( x\right) \) . By definition, we have \( \zeta = \mathrm{w} - \lim {\zeta }_{i} \), where \( {\zeta }_{i} \in {\partial }_{P}\left( {{f}_{1} + {f}_{2}}\right) \left( {x}_{i}\right) \) and where \( {x}_{i} \rightarrow x,\left( {{f}_{1} + {f}_{2}}\right) \left( {x}_{i}\right) \rightarrow \left( {{f}_{1} + {f}_{2}}\right) \left( x\right) \) . If the fuzzy Sum Rule (Theorem 8.2) can be applied, for example if one of \( {f}_{1},{f}_{2} \) is Lipschitz locally, then we can write \[ {\zeta }_{i} = {\theta }_{i} + {\xi }_{i} + {\varepsilon }_{i}{u}_{i} \] (4) where \( {\theta }_{i} \in {\partial }_{P}{f}_{1}\left( {x}_{i}^{\prime }\right) ,{\xi }_{i} \in {\partial }_{P}{f}_{2}\left( {x}_{i}^{\prime \prime }\right) ,{u}_{i} \in B \), where \( {x}_{i}^{\prime } \) and \( {x}_{i}^{\prime \prime } \) lie in an \( {\varepsilon }_{i} \) -ball around \( {x}_{i} \), and where \( {\varepsilon }_{i} \downarrow 0 \) as \( i \rightarrow \infty \) . We also have that \( {f}_{1}\left( {x}_{i}^{\prime }\right) \) and \( {f}_{2}\left( {x}_{i}^{\prime \prime }\right) \) are \( {\varepsilon }_{i} \) -close to \( {f}_{1}\left( {x}_{i}\right) \) and \( {f}_{2}\left( {x}_{i}\right) \), respectively. We now wish to pass to the limit in (4); How can this be carried out? Again, assuming that one of the functions (let us say \( {f}_{1} \) ) is Lipschitz locally would do the trick. For in that case \( {\partial }_{P}{f}_{1}\left( \cdot \right) \) is bounded locally (by the Lipschitz rank) so that the sequence \( \left\{ {\theta }_{i}\right\} \) is bounded. But \( \left\{ {\zeta }_{i}\right\} \) is bounded too, since this sequence converges weakly (to \( \zeta \) ). It follows from (4) that \( \left\{ {\xi }_{i}\right\} \) is bounded. Extracting subsequences so that all these sequences converge weakly (to \( \theta ,\zeta ,\xi \), respectively) gives \( \zeta = \theta + \xi \) . We also have that \( \left\{ {{f}_{1}\left( {x}_{i}^{\prime }\right) }\right\} \) and \( \left\{ {{f}_{2}\left( {x}_{i}^{\prime \prime }\right) }\right\} \) converge to \( {f}_{1}\left( x\right) \) and \( {f}_{2}\left( x\right) \), respectively (why?), so that \( \theta \in {\partial }_{L}{f}_{1}\left( x\right) ,\xi \in {\partial }_{L}{f}_{2}\left( x\right) \) by definition. The discussion has proved 10.1. Proposition. If one of \( {f}_{1},{f}_{2} \) is Lipschitz near \( x \), then \[ {\partial }_{L}\left( {{f}_{1} + {f}_{2}}\right) \left( x\right) \subseteq {\partial }_{L}{f}_{1}\left( x\right) + {\partial }_{L}{f}_{2}\left( x\right) . \] This is an appealing (nonfuzzy) Sum Rule. In view of the inclusion (1), we might be led to believe (briefly) that \( {\partial }_{L}{f}_{1}\left( x\right) + {\partial }_{L}{f}_{2}\left( x\right) \subseteq {\partial }_{L}\left( {{f}_{1} + {f}_{2}}\right) \left( x\right) \) , and hence that equality actually holds in Proposition 10.1. But alas, the limiting form of (1) fails to hold, as shown in the next exercise. (We will find supplementary hypotheses in Chapter 2 under which equality does hold.) 10.2. Exercise. Set \( X = \mathbb{R},{f}_{1}\left( x\right) \mathrel{\text{:=}} \left| x\right| ,{f}_{2}\left( x\right) \mathrel{\text{:=}} - \left| x\right| \) . Show that \( {\partial }_{L}{f}_{1}\left( 0\right) = \left\lbrack {-1,1}\right\rbrack ,{\partial }_{L}{f}_{2}\left( 0\right) = \{ - 1,1\} \) . Verify that the conclusion of Proposition 10.1 holds, but not with equality. Let us consider now the limiting issue in geometric terms. The natural closure operation to apply to \( {N}_{S}^{P}\left( \cdot \right) \) gives rise to \( {N}_{S}^{L}\left( x\right) \), the limiting normal (or \( L \) -normal) cone to \( S \) at \( x \in S \) : \[ {N}_{S}^{L}\left( x\right) \mathrel{\text{:=}} \left\{ {\mathrm{w} - \lim {\zeta }_{i} : {\zeta }_{i} \in {N}_{S}^{P}\left( {x}_{i}\right) ,{x}_{i}\overset{S}{ \rightarrow }x}\right\} . \] Here, \( {x}_{i}\overset{S}{ \rightarrow }x \) signifies that \( {x}_{i} \rightarrow x \) and that \( {x}_{i} \in S\forall i \) . One motivation for defining a limiting normal cone is that \( {N}_{S}^{P}\left( x\right) \) is potentially trivial (i.e., \( = \{ 0\} ) \) for "many" \( x \) ; in pointwise considerations it is \( {N}_{S}^{L}\left( x\right) \) that may incorporate normality information. The following exercise vindicates this hope in \( {\mathbb{R}}^{n} \), confirms a certain coherence between the functional and geometric closure operations, and derives some basic properties of the limiting constructs. 10.3. Exercise. Let \(
1012_(GTM178)Nonsmooth Analysis and Control Theory
22
\( {N}_{S}^{P}\left( \cdot \right) \) gives rise to \( {N}_{S}^{L}\left( x\right) \), the limiting normal (or \( L \) -normal) cone to \( S \) at \( x \in S \) : \[ {N}_{S}^{L}\left( x\right) \mathrel{\text{:=}} \left\{ {\mathrm{w} - \lim {\zeta }_{i} : {\zeta }_{i} \in {N}_{S}^{P}\left( {x}_{i}\right) ,{x}_{i}\overset{S}{ \rightarrow }x}\right\} . \] Here, \( {x}_{i}\overset{S}{ \rightarrow }x \) signifies that \( {x}_{i} \rightarrow x \) and that \( {x}_{i} \in S\forall i \) . One motivation for defining a limiting normal cone is that \( {N}_{S}^{P}\left( x\right) \) is potentially trivial (i.e., \( = \{ 0\} ) \) for "many" \( x \) ; in pointwise considerations it is \( {N}_{S}^{L}\left( x\right) \) that may incorporate normality information. The following exercise vindicates this hope in \( {\mathbb{R}}^{n} \), confirms a certain coherence between the functional and geometric closure operations, and derives some basic properties of the limiting constructs. 10.3. Exercise. Let \( S \) be closed and \( f \in \mathcal{F} \) . (a) \( \left( {\zeta , - 1}\right) \in {N}_{\text{epi }f}^{L}\left( {x, f\left( x\right) }\right) \) iff \( \zeta \in {\partial }_{L}f\left( x\right) \), and \( {\partial }_{L}{I}_{S}\left( x\right) = {N}_{S}^{L}\left( x\right) \) . (b) \( {N}_{S}^{L}\left( x\right) = {N}_{S}^{P}\left( x\right) \) when \( S \) is convex. (c) When \( X = {\mathbb{R}}^{n} \), both \( {N}_{S}^{L}\left( x\right) \) and \( {\partial }_{L}f\left( x\right) \) are closed sets; if \( x \in \operatorname{bdry}S \), then \( {N}_{S}^{L}\left( x\right) \neq \{ 0\} \), and if \( f \) is Lipschitz near \( x \) , then \( {\partial }_{L}f\left( x\right) \neq \varnothing \) . Although a type of closure operation was used in defining \( {N}_{S}^{L}\left( x\right) \), it is a fact that this set may fail to be closed when \( X \) is infinite dimensional; similarly, \( {\partial }_{L}f\left( x\right) \) may not be closed if \( f \) fails to be Lipschitz. These facts make the limiting calculus most appealing in the presence of Lipschitz hypotheses, or in finite dimensions. Here is a limiting form of the Chain Rule in such a context, one whose use will be illustrated in the end-of-chapter problems. 10.4. Theorem. Let \( F : X \rightarrow {\mathbb{R}}^{n} \) be Lipschitz near \( x \), and let \( g : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be Lipschitz near \( F\left( x\right) \) . Then \[ {\partial }_{L}\left( {g \circ F}\right) \left( x\right) \subseteq \left\{ {{\partial }_{L}\langle \gamma, F\left( \cdot \right) \rangle \left( x\right) : \gamma \in {\partial }_{L}g\left( {F\left( x\right) }\right) }\right\} . \] ## 11 Problems on Chapter 1 11.1. Give examples of each the following: (a) A set \( S \) in \( {\mathbb{R}}^{2} \) and a point \( s \in S \) such that \( {N}_{S}^{P}\left( s\right) \) is neither open nor closed, and a function \( f : \mathbb{R} \rightarrow \mathbb{R} \) with a point \( x \in \operatorname{dom}f \) such that \( {\partial }_{P}f\left( x\right) \) is neither open nor closed. (b) A function \( f \in \mathcal{F} \) that is not bounded below on \( \bar{B} \) . (c) A function \( f \in \mathcal{F} \) that is bounded below on \( \bar{B} \) but does not attain its minimum there. 11.2. Suppose that \( {\left\{ {e}_{i}\right\} }_{i = 1}^{\infty } \) is an orthonormal basis for \( X \), and set \( S \mathrel{\text{:=}} \) \( \overline{\operatorname{co}}\left\{ {\pm {e}_{i}/i : i \geq 1}\right\} \) . Prove that \( {N}_{S}^{P}\left( 0\right) = \{ 0\} \) . 11.3. Suppose \( S \subset X \) is such that \( \forall x \in S,\forall \zeta \in {N}_{S}^{P}\left( x\right) \), we have \[ \left\langle {\zeta ,{x}^{\prime } - x}\right\rangle \leq 0\forall {x}^{\prime } \in S. \] Prove that \( S \) is convex. (This provides a converse to Proposition 1.10(a).) 11.4. Let \( S \) be bounded, closed, and nonempty. Prove that \( \mathop{\bigcup }\limits_{{x \in S}}{N}_{S}^{P}\left( x\right) \) is dense in \( X \) . 11.5. Let \( s \) belong to a closed set \( S \), and suppose that \( {d}_{S} \) is differentiable at \( s \) . Show that \( {d}_{S}^{\prime }\left( s\right) = 0 \) necessarily. Prove that if \( {N}_{S}^{P}\left( s\right) \neq \{ 0\} \), then \( {d}_{S} \) cannot be differentiable at \( s \) . 11.6. Let \( S \subseteq X \) be nonempty (but not necessarily closed), and suppose \( f \in \mathcal{F} \) satisfies \( f\left( x\right) \geq - c\parallel x\parallel \) for some \( c > 0 \) . (a) Let \( K > c \), and define the function \( g : X \rightarrow \mathbb{R} \) by \[ g\left( x\right) \mathrel{\text{:=}} \mathop{\inf }\limits_{{s \in S}}\{ f\left( s\right) + K\parallel s - x\parallel \} . \] Show that \( g \) is Lipschitz on \( X \) of rank \( K \) . (b) Suppose in addition that \( f \) is Lipschitz of rank \( K \) on \( S \) . Prove that \( g \) agrees with \( f \) on \( S \), and deduce that Lipschitz functions defined on any subset of \( X \) can be extended to a Lipschitz function on all of \( X \) without increasing the Lipschitz rank. 11.7. Let \( f \in \mathcal{F} \) and let \( M \in \mathbb{R} \) . Set \[ {f}^{M}\left( x\right) \mathrel{\text{:=}} \min \{ f\left( x\right), M\} . \] (a) Show that \( {f}^{M} \in \mathcal{F} \) . (b) If \( \zeta \in {\partial }_{P}{f}^{M}\left( x\right) \), prove that \( \zeta \in {\partial }_{P}f\left( x\right) \) if \( f\left( x\right) \leq M \), and that \( \zeta = 0 \) otherwise. 11.8. Suppose that \( L : {\mathbb{R}}^{n} \rightarrow ( - \infty ,\infty \rbrack \) is lower semicontinuous, not identically \( + \infty \), and bounded below. For any \( v\left( \cdot \right) \in X \mathrel{\text{:=}} {L}_{n}^{2}\left\lbrack {0,1}\right\rbrack \), define \( f : X \rightarrow ( - \infty ,\infty \rbrack \) by \[ f\left( {v\left( \cdot \right) }\right) \mathrel{\text{:=}} {\int }_{0}^{1}L\left( {v\left( t\right) }\right) {dt} \] Prove that \( f \) is lower semicontinuous, and that for any \( \alpha > 0 \), the function \[ F\left( {u\left( \cdot \right) }\right) \mathrel{\text{:=}} {\int }_{0}^{1}\left\{ {L\left( {u\left( t\right) }\right) + \alpha \parallel v\left( t\right) - u\left( t\right) {\parallel }^{2}}\right\} {dt} \] attains a unique minimum over \( X \) for a dense set of \( v\left( \cdot \right) \in X \) . 11.9. Suppose \( f \in \mathcal{F} \) is bounded below, \( \alpha > 0 \), and \( {f}_{\alpha } \) is defined as in \( §5 \) . (a) Let \( x \in \operatorname{dom}f \) . Prove that for each \( \alpha > 0 \) there exists \( {r}_{\alpha } > 0 \) so that each minimizer \( \bar{y} \) in the infimum defining \( {f}_{\alpha }\left( x\right) \) satisfies \( \bar{y} \in B\left( {x;{r}_{\alpha }}\right) \) , and \( {r}_{\alpha } \rightarrow 0 \) as \( \alpha \rightarrow \infty \) . (b) Prove that \( {f}_{\alpha } \uparrow f\left( x\right) \) as \( \alpha \uparrow \infty \) for each \( x \in X \), and that for \( x \in \operatorname{dom}f \) , \[ \mathop{\liminf }\limits_{{y \rightarrow x,\alpha \rightarrow \infty }}{f}_{a}\left( y\right) \geq f\left( x\right) \] (Thus \( {f}_{\alpha } \) is a locally Lipschitz lower approximation to \( f \), one that improves as \( \alpha \rightarrow \infty \) ; it is called the Moreau-Yosida approximation.) 11.10. Suppose \( f \in \mathcal{F} \) is bounded below. Show that there exist sequences \( \left\{ {y}_{i}\right\} \) and \( \left\{ {\zeta }_{i}\right\} \) so that \( {\zeta }_{i} \in {\partial }_{P}f\left( {y}_{i}\right) \) for each \( i \), and satisfying \( f\left( {y}_{i}\right) \rightarrow \) \( \mathop{\inf }\limits_{{x \in X}}f\left( x\right) \) and \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \rightarrow 0 \) as \( i \rightarrow \infty \) . 11.11. Let \( f \in \mathcal{F} \) and \( {x}_{0} \in \operatorname{dom}f \) . Suppose that for some \( \varepsilon > 0 \), for all \( x \in {x}_{0} + {\varepsilon B} \) satisfying \( \left| {f\left( x\right) - f\left( {x}_{0}\right) }\right| < \varepsilon \), for all \( \zeta \in {\partial }_{P}f\left( x\right) \), we have \( \parallel \zeta \parallel \leq K \) . Prove that \( f \) is Lipschitz of rank \( K \) on a neighborhood of \( {x}_{0} \) . (Hint. Consider \( \widetilde{f}\left( x\right) \mathrel{\text{:=}} \min \left\{ {f\left( x\right), f\left( {x}_{0}\right) + \varepsilon /2}\right\} \) in light of Problem 11.7.) 11.12. With \( X = \mathbb{R} \), calculate \( {\partial }_{P}f\left( 0\right) \) and \( {\partial }_{L}f\left( 0\right) \) when \( f\left( x\right) \mathrel{\text{:=}} {x}^{2}\sin \left( {1/x}\right) \) . 11.13. (a) Prove the following Proximal Mean Value Theorem: Theorem. Let \( f \in \mathcal{F}\left( X\right) \) be locally Lipschitz on a neighborhood of the line segment \( \left\lbrack {x, y}\right\rbrack \) . Then \( \forall \varepsilon > 0 \) there exists a point \( z \) in the \( \varepsilon \) -neighborhood of \( \left\lbrack {x, y}\right\rbrack \) and \( \zeta \in {\partial }_{P}f\left( z\right) \) such that \[ f\left( y\right) - f\left( x\right) \leq \langle \zeta, y - x\rangle + \varepsilon . \] (b) Show that the conclusion is false in the form \[ \left| {f\left( y\right) - f\left( x\right) -\langle \zeta, y - x\rangle }\right| < \varepsilon . \] (c) Obtain a limiting form of the theorem in terms of \( {\partial }_{L}f \) when \( \varepsilon \rightarrow 0 \) . (Hint. Minimize \( \varphi \left( {t, u}\right) \mathrel{\text{:=}} f\left( {x + u}\right) - {tf}\left( y\right) - \left( {1 - t}\right) f\left( x\right) \) over the set of \( \left( {t, u}\right) \) satisfying \( u = t\left( {y - x}\right) ,0 \leq t < 1 \) .) 11.14. Let \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be defined by \[ f\left( x\right) = f\left( {{x}_{1},{x}_{2},\ldots ,{x}_{n}}\right) = \mathop{\max }\limits_{{1 \leq i \leq n}}{x}_{i}. \] Let \( M\left( x\right) \) be the set of those indices \( i \) at which this maximum is attained. Prove that \( {\partial }_{P}f\left( x\right) \) consists of all those vectors \( \left( {{\zeta }_{1},{\zeta }_{
1012_(GTM178)Nonsmooth Analysis and Control Theory
23
eq \langle \zeta, y - x\rangle + \varepsilon . \] (b) Show that the conclusion is false in the form \[ \left| {f\left( y\right) - f\left( x\right) -\langle \zeta, y - x\rangle }\right| < \varepsilon . \] (c) Obtain a limiting form of the theorem in terms of \( {\partial }_{L}f \) when \( \varepsilon \rightarrow 0 \) . (Hint. Minimize \( \varphi \left( {t, u}\right) \mathrel{\text{:=}} f\left( {x + u}\right) - {tf}\left( y\right) - \left( {1 - t}\right) f\left( x\right) \) over the set of \( \left( {t, u}\right) \) satisfying \( u = t\left( {y - x}\right) ,0 \leq t < 1 \) .) 11.14. Let \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be defined by \[ f\left( x\right) = f\left( {{x}_{1},{x}_{2},\ldots ,{x}_{n}}\right) = \mathop{\max }\limits_{{1 \leq i \leq n}}{x}_{i}. \] Let \( M\left( x\right) \) be the set of those indices \( i \) at which this maximum is attained. Prove that \( {\partial }_{P}f\left( x\right) \) consists of all those vectors \( \left( {{\zeta }_{1},{\zeta }_{2},\ldots ,{\zeta }_{n}}\right) \) such that \( {\zeta }_{i} \geq 0,\mathop{\sum }\limits_{{i = 1}}^{n}{\zeta }_{i} = 1,{\zeta }_{i} = 0 \) if \( i \notin M\left( x\right) \) . (Hint. \( f \) is convex.) 11.15. (Monotonicity.) Let \( C \) be a cone in \( X \), and let \( f \in \mathcal{F}\left( X\right) \) . We say that \( f \) is \( C \) -decreasing if \( y - x \in C \) implies \( f\left( y\right) \leq f\left( x\right) \) . Prove that if \( f \) is locally Lipschitz, then \( f \) is \( C \) -decreasing iff \( \zeta \in {C}^{ \circ }\forall \zeta \in {\partial }_{P}f\left( x\right) ,\forall x \in X \) , where the polar \( {C}^{ \circ } \) of \( C \) is defined as \[ {C}^{ \circ } \mathrel{\text{:=}} \{ z \in X : \langle z, c\rangle \leq 0\forall c \in C\} . \] 11.16. Prove the Limiting Chain Rule, Theorem 10.4. 11.17. (Upper envelopes.) Let \( {f}_{i} \in \mathcal{F}\left( X\right) \) be locally Lipschitz, \( i = 1,2,\ldots, n \) , and set \( f\left( x\right) \mathrel{\text{:=}} \mathop{\max }\limits_{{1 \leq i \leq n}}{f}_{i}\left( x\right) \) . We denote by \( M\left( x\right) \) the set of those indices \( i \) at which the maximum defining \( f\left( x\right) \) is attained. (a) Prove that \( f \) is locally Lipschitz, and that \[ \operatorname{co}\mathop{\bigcup }\limits_{{i \in M\left( x\right) }}{\partial }_{P}{f}_{i}\left( x\right) \subset {\partial }_{P}f\left( x\right) \] (b) If \( \zeta \in {\partial }_{L}f\left( x\right) \), then there exist \( {\gamma }_{i} \geq 0\left( {i = 1,2,\ldots, n}\right) \) with \( \mathop{\sum }\limits_{{i = 1}}^{n}{\gamma }_{i} = \) \( 1 \) and \( {\gamma }_{i} = 0 \) for \( i \notin M\left( x\right) \) such that \( \zeta \in {\partial }_{L}\left( {\mathop{\sum }\limits_{{i = 1}}^{n}{\gamma }_{i}{f}_{i}}\right) \left( x\right) \) . 11.18. (The Lagrange Multiplier Rule for inequality constraints.) Let \( {x}_{0} \) solve the problem of minimizing \( {g}_{0}\left( x\right) \) subject to \( {g}_{i}\left( x\right) \leq 0\left( {i = 1,2,\ldots, n}\right) \) , where all the functions involved are locally Lipschitz. (a) Show that \( {x}_{0} \) minimizes (subject to no constraints) the function \[ x \mapsto \max \left\{ {{g}_{0}\left( x\right) - {g}_{0}\left( {x}_{0}\right) ,{g}_{1}\left( x\right) ,\ldots ,{g}_{n}\left( x\right) }\right\} . \] (b) Deduce the existence of \( {\gamma }_{i} \geq 0\left( {i = 0,1,\ldots, n}\right) \) with \( \mathop{\sum }\limits_{{i = 0}}^{n}{\gamma }_{i} = 1 \) and \( {\gamma }_{i}{g}_{i}\left( {x}_{0}\right) = 0 \) for \( i \geq 1 \), such that \( 0 \in {\partial }_{L}\left( {\mathop{\sum }\limits_{{i = 0}}^{n}{\gamma }_{i}{g}_{i}}\right) \left( {x}_{0}\right) \) . 11.19. (Partial subgradients.) Let \( f \) be locally Lipschitz on \( {X}_{1} \times {X}_{2} \), and let \( {\bar{\zeta }}_{1} \in {\partial }_{P}f\left( {\cdot ,{\bar{x}}_{2}}\right) \left( {\bar{x}}_{1}\right) \) (i.e., \( {\bar{\zeta }}_{1} \) is a proximal subgradient at \( {\bar{x}}_{1} \) of the function \( \left. {{x}_{1} \mapsto f\left( {{x}_{1},{\bar{x}}_{2}}\right) }\right) \) . Then for any \( \varepsilon > 0 \) there exist \( \left( {{\widetilde{x}}_{1},{\widetilde{x}}_{2}}\right) \) in \( \left( {{\bar{x}}_{1},{\bar{x}}_{2}}\right) + {\varepsilon B} \) and \( \left( {{\widetilde{\zeta }}_{1},{\widetilde{\zeta }}_{2}}\right) \) in \( {\partial }_{P}f\left( {{\widetilde{x}}_{1},{\widetilde{x}}_{2}}\right) \) such that \( \begin{Vmatrix}{{\widetilde{\zeta }}_{1} - {\bar{\zeta }}_{1}}\end{Vmatrix} < \varepsilon \) . (Hint. Take \( F\left( x\right) \mathrel{\text{:=}} \left( {x,{\bar{x}}_{2}}\right) \) in Theorem 9.1.) 11.20. This exercise is devoted to some relationships between Fréchet and Gâteaux differentiability. (a) Suppose \( f : X \rightarrow \mathbb{R} \) is Fréchet differentiable at \( x \) . Show that \( f \) is continuous at \( x \) . (b) Suppose \( f : X \rightarrow \mathbb{R} \) is Gâteaux differentiable at \( x \) . Show that for each \( v \in X \), the function \( g\left( t\right) \mathrel{\text{:=}} f\left( {x + {tv}}\right) \) defined for small values of \( \left| t\right| \) is continuous at \( t = 0 \) . (c) Consider the following functions: \[ {f}_{1}\left( {x, y}\right) = \left\{ {\begin{array}{ll} \frac{{y}^{2}}{x} & \text{ if }x \neq 0, \\ 0 & \text{ if }x = 0, \end{array}\;{f}_{2}\left( {x, y}\right) = \left\{ \begin{array}{ll} \frac{xy}{{x}^{2} + {y}^{2}} & \text{ if }\left( {x, y}\right) \neq \left( {0,0}\right) , \\ 0 & \text{ if }\left( {x, y}\right) = \left( {0,0}\right) . \end{array}\right. }\right. \] Show that \( {f}_{1} \) has a Gâteaux derivative at the origin, but that \( {f}_{1} \) is not continuous there (and therefore not Fréchet differentiable). Show that the partial derivatives of \( {f}_{2} \) exist at \( \left( {0,0}\right) \), but the Gâteaux derivative \( {f}_{2}^{\prime }\left( {0,0}\right) \) does not. (d) Suppose that \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is Lipschitz near \( x \) . Then at \( x \), Gâteaux differentiability implies Fréchet differentiability (this is false in general in infinite dimensional spaces). (e) With \( X \) a Hilbert space, show that \( {f}^{\prime }\left( x\right) \) exists iff we have, for some \( \zeta \in X \) \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{f\left( {x + {t}_{i}{v}_{i}}\right) - f\left( x\right) }{{t}_{i}} = \langle \zeta, v\rangle \forall x \] whenever \( {t}_{i} \) decreases to 0 and \( {v}_{i} \) converges weakly to \( v \) (and then \( \left. {{f}^{\prime }\left( x\right) = \zeta }\right) \) . (f) Suppose \( f \) has a Gâteaux derivative at each point near \( x \in \mathcal{F} \), and \( y \mapsto {f}^{\prime }\left( y\right) \) is continuous at \( x \) . Show that \( f \) is Fréchet differentiable at \( x \) . 11.21. Let \( X \) be separable, and let \( f \in \mathcal{F}\left( X\right) \) be Lipschitz near a point \( \bar{x} \), let \( \left\{ {x}_{i}\right\} \) be a sequence converging to \( \bar{x} \), and let \( \left\{ {\zeta }_{i}\right\} \) be a sequence converging weakly to \( \zeta \), where \( {\zeta }_{i} \in {\partial }_{L}f\left( {x}_{i}\right) \) for each \( i \) . Prove that \( \zeta \in {\partial }_{L}f\left( \bar{x}\right) \) . (Hint. Recall that the weak topology restricted to a closed bounded subset of \( X \) is equivalent to a metric one.) 11.22. Let \( X \) admit an orthonormal basis \( {\left\{ {e}_{i}\right\} }_{i = 1}^{\infty } \), and let \( f \in \mathcal{F}\left( X\right) \) be such that at every \( x \in \operatorname{dom}f \) we have, for every index \( i \), the conditions \[ \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{f\left( {x + t{e}_{i}}\right) - f\left( x\right) }{t} \leq 0,\;\mathop{\liminf }\limits_{{t \downarrow 0}}\frac{f\left( {x - t{e}_{i}}\right) - f\left( x\right) }{t} \leq 0. \] Prove that \( f \) is a constant function. 11.23. (Horizontal Approximation Theorem.) This problem (needed in Chapter 4) will prove that horizontal proximal normals to epigraphs can be approximated by nonhorizontal ones (which then correspond to subgradients). Formally: Theorem. Let \( f \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \), and let \( \left( {\theta ,0}\right) \in {N}_{\text{epi }f}^{P}\left( {x, f\left( x\right) }\right) \) . Then for every \( \varepsilon > 0 \) there exist \( {x}^{\prime } \in x + {\varepsilon B} \) and \( \left( {\zeta , - \lambda }\right) \in {N}_{\text{epi }f}^{P}\left( {{x}^{\prime }, f\left( {x}^{\prime }\right) }\right) \) such that \[ \lambda > 0,\;\left| {f\left( {x}^{\prime }\right) - f\left( x\right) }\right| < \varepsilon ,\;\parallel \left( {\theta ,0}\right) - \left( {\zeta , - \lambda }\right) \parallel < \varepsilon . \] (a) Show how the case \( \theta = 0 \) is an immediate consequence of the Density Theorem 3.1. We assume henceforth that \( x = 0, f\left( 0\right) = 0,\theta \neq 0 \), and that the point \( \left( {\theta ,0}\right) \) admits unique closest point \( \left( {0,0}\right) \) in epi \( f = : S \) . (Why does this not constitute a loss of generality?) (b) We have \( {d}_{S}\left( {\theta, t}\right) > {d}_{S}\left( {\theta ,0}\right) = \parallel \theta \parallel \forall t < 0 \) . (c) The function \( t \mapsto {d}_{S}\left( {\theta, t}\right) \) has strictly negative proximal subgradients at points \( t < 0 \) arbitrarily near 0 . (d) There exist points \( \left( {x, t}\right) \) arbitrarily near \( \left( {\theta ,0}\right) \) with \( t < 0 \) such that an (in fact, the only) element of \( {\partial }_{P}{d}_{S}\left( {x, t}\right) \) has strictly negative final component. (Hint. Problem 11.19.) (e) As the points \( \left( {x, t}\right) \) of part (d) converge to \( \left( {\theta ,0}\right) \), the corresponding element of \( {\partial }_{P}{d}_{S}\left( {x, t}\right) \) converges to \( \left( {\theta ,0}\right) \), completing the proof. 11.24. Let \( f : X \rightarrow \mathbb{R} \) be continuous and convex. Prove that \( f \) is Fréchet differentiable at a
1012_(GTM178)Nonsmooth Analysis and Control Theory
24
\( f = : S \) . (Why does this not constitute a loss of generality?) (b) We have \( {d}_{S}\left( {\theta, t}\right) > {d}_{S}\left( {\theta ,0}\right) = \parallel \theta \parallel \forall t < 0 \) . (c) The function \( t \mapsto {d}_{S}\left( {\theta, t}\right) \) has strictly negative proximal subgradients at points \( t < 0 \) arbitrarily near 0 . (d) There exist points \( \left( {x, t}\right) \) arbitrarily near \( \left( {\theta ,0}\right) \) with \( t < 0 \) such that an (in fact, the only) element of \( {\partial }_{P}{d}_{S}\left( {x, t}\right) \) has strictly negative final component. (Hint. Problem 11.19.) (e) As the points \( \left( {x, t}\right) \) of part (d) converge to \( \left( {\theta ,0}\right) \), the corresponding element of \( {\partial }_{P}{d}_{S}\left( {x, t}\right) \) converges to \( \left( {\theta ,0}\right) \), completing the proof. 11.24. Let \( f : X \rightarrow \mathbb{R} \) be continuous and convex. Prove that \( f \) is Fréchet differentiable at a dense set of points. (Hint. Recall Exercise 2.8.) 11.25. Let \( f \) be locally Lipschitz on \( X \) . (a) Prove that \( {\partial }_{L}f\left( x\right) \) reducing to a singleton is a sufficient but not a necessary condition for \( {f}^{\prime }\left( x\right) \) to exist. (b) When \( X \) is finite dimensional, prove that \( f \) is \( {C}^{1} \) on an open set \( U \) iff \( {\partial }_{L}f\left( x\right) \) reduces to a singleton for every \( x \in U \) . 11.26. Let \( f \in \mathcal{F}\left( X\right) \), and set \( S \mathrel{\text{:=}} \{ x \in X : f\left( x\right) \leq 0\} \) . Let \( \bar{x} \) satisfy \( f\left( \bar{x}\right) = 0 \) . Then (a) \( \mathop{\bigcup }\limits_{{\lambda \geq 0}}\lambda {\partial }_{P}f\left( \bar{x}\right) \subset {N}_{S}^{P}\left( \bar{x}\right) \) . (b) If \( f \) is Lipschitz near \( \bar{x} \) and \( 0 \notin {\partial }_{L}f\left( \bar{x}\right) \), then we have \[ {N}_{S}^{L}\left( \bar{x}\right) \subset \mathop{\bigcup }\limits_{{\lambda \geq 0}}\lambda {\partial }_{L}f\left( \bar{x}\right) \] (Hint. One approach uses Problems 11.18 and 11.25.) (c) If \( f \) is \( {C}^{2} \) near \( \bar{x} \) and \( {f}^{\prime }\left( \bar{x}\right) \neq 0 \), then \[ {N}_{S}^{P}\left( \bar{x}\right) = {N}_{S}^{L}\left( \bar{x}\right) = \left\{ {\lambda {f}^{\prime }\left( \bar{x}\right) : \lambda \geq 0}\right\} . \] 11.27. Let \( S \) be a nonempty closed subset of \( X,\bar{x} \in S \) . We will prove the formula \[ {N}_{S}^{L}\left( \bar{x}\right) = \mathop{\bigcup }\limits_{{\lambda \geq 0}}\lambda {\partial }_{L}{d}_{S}\left( \bar{x}\right) \] (a) Prove that if \( \zeta \in {N}_{S}^{P}\left( x\right) \), then for any \( \varepsilon > 0 \) we have \( \zeta /\left( {\parallel \zeta \parallel + \varepsilon }\right) \in \) \( {\partial }_{P}{d}_{S}\left( x\right) \) . (Hint. Use exact penalization.) (b) Deduce that \( {N}_{S}^{L}\left( \bar{x}\right) \subset \mathop{\bigcup }\limits_{{\lambda \geq 0}}\lambda {\partial }_{L}{d}_{S}\left( \bar{x}\right) \) . (c) Complete the proof by confirming the opposite inclusion. 11.28. Let \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be locally Lipschitz. Prove that \( {\partial }_{L}f \) is countably generated, in the following sense. There exists a countable set \( \Omega \) in \( {\mathbb{R}}^{n} \times {\mathbb{R}}^{n} \) such that for every \( x \), we have \[ {\partial }_{L}f\left( x\right) = \left\{ {\mathop{\lim }\limits_{{i \rightarrow \infty }}{\zeta }_{i} : \left( {{x}_{i},{\zeta }_{i}}\right) \in \Omega \forall i,\mathop{\lim }\limits_{{i \rightarrow \infty }}{x}_{i} = x}\right\} . \] (Hint. Consider a countable dense subset of \( \operatorname{gr}{\partial }_{P}f \) .) 11.29. (Proximal Gronwall Inequality.) Let \( f : \left\lbrack {0, T}\right\rbrack \rightarrow \mathbb{R} \) be locally Lipschitz, where \( T > 0 \), and suppose there exists a constant \( M \geq 0 \) such that \[ \zeta \leq {Mf}\left( t\right) \forall \zeta \in {\partial }_{P}f\left( t\right) ,\forall t \in \left( {0, T}\right) . \] Then \[ f\left( t\right) \leq {e}^{Mt}f\left( 0\right) \forall t \in \left\lbrack {0, T}\right\rbrack . \] 11.30. Let \( f \in \mathcal{F} \) and \( U \subset X \) be open. Show that \( f \) is \( {C}^{1} \) on \( U \) iff for all \( x \in U \) and \( \varepsilon > 0 \) there exists \( \delta > 0 \) such that \[ \begin{Vmatrix}{{x}_{i} - x}\end{Vmatrix} < \delta ,\;{\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) \;\left( {i = 1,2}\right) \] implies that \( \begin{Vmatrix}{{\zeta }_{1} - {\zeta }_{2}}\end{Vmatrix} < \varepsilon \) . 2 Generalized Gradients in Banach Space I drew up the state of my affairs in writing; not so much to leave them to any that were to come after me, for I was like to have but few heirs, as to deliver my thoughts from daily poring upon them, and afflicting my mind. -Daniel Defoe, Robinson Crusoe The calculus of generalized gradients is the best-known and most frequently invoked part of nonsmooth analysis. Unlike proximal calculus, it can be developed in an arbitrary Banach space \( X \) . In this chapter we make a fresh start in such a setting, but this time, in contrast to Chapter 1, we begin with functions and not sets. We present the basic results for the class of locally Lipschitz functions. Then the associated geometric concepts are introduced, including for the first time a look at tangency. In fact, we examine two notions of tangency; sets for which they coincide are termed regular and enjoy useful properties. We proceed to relate the generalized gradient to the constructs of the preceding chapter when \( X \) is a Hilbert space. Finally, we derive a useful limiting-gradient characterization when the underlying space is finite dimensional. ## 1 Definition and Basic Properties Throughout this chapter, \( X \) is a real Banach space. Let \( f : X \rightarrow \mathbb{R} \) be Lipschitz of rank \( K \) near a given point \( x \in X \) ; that is, for some \( \varepsilon > 0 \), we have \[ \left| {f\left( y\right) - f\left( z\right) }\right| \leq K\parallel y - z\parallel \forall y, z \in B\left( {x;\varepsilon }\right) . \] The generalized directional derivative of \( f \) at \( x \) in the direction \( v \), denoted \( {f}^{ \circ }\left( {x;v}\right) \), is defined as follows: \[ {f}^{ \circ }\left( {x;v}\right) \mathrel{\text{:=}} \mathop{\limsup }\limits_{\substack{{y \rightarrow x} \\ {t \downarrow 0} }}\frac{f\left( {y + {tv}}\right) - f\left( y\right) }{t}, \] where of course \( y \) is a vector in \( X \) and \( t \) is a positive scalar. Note that this definition does not presuppose the existence of any limit (since it involves an upper limit only), that it involves only the behavior of \( f \) arbitrarily near \( x \), and that it differs from the traditional definition of the directional derivative in that the base point \( \left( y\right) \) of the difference quotient varies. The utility of \( {f}^{ \circ } \) stems from the following basic properties. (A function \( g \) is positively homogeneous if \( g\left( {\lambda v}\right) = {\lambda g}\left( v\right) \) for \( \lambda \geq 0 \), and subadditive if \( g\left( {v + w}\right) \leq g\left( v\right) + g\left( w\right) \) .) 1.1. Proposition. Let \( f \) be Lipschitz of rank \( K \) near \( x \) . Then: (a) The function \( v \mapsto {f}^{ \circ }\left( {x;v}\right) \) is finite, positively homogeneous, and subadditive on \( X \), and satisfies \[ \left| {{f}^{ \circ }\left( {x;v}\right) }\right| \leq K\parallel v\parallel \] (b) \( {f}^{ \circ }\left( {x;v}\right) \) is upper semicontinuous as a function of \( \left( {x, v}\right) \) and, as a function of \( v \) alone, is Lipschitz of rank \( K \) on \( X \) . (c) \( {f}^{ \circ }\left( {x; - v}\right) = {\left( -f\right) }^{ \circ }\left( {x;v}\right) \) . Proof. In view of the Lipschitz condition, the absolute value of the difference quotient in the definition of \( {f}^{ \circ }\left( {x;v}\right) \) is bounded by \( K\parallel v\parallel \) when \( y \) is sufficiently near \( x \) and \( t \) sufficiently near 0 . It follows that \( \left| {{f}^{ \circ }\left( {x;v}\right) }\right| \) admits the same upper bound. The fact that \( {f}^{ \circ }\left( {x;{\lambda v}}\right) = \lambda {f}^{ \circ }\left( {x;v}\right) \) for any \( \lambda \geq 0 \) is immediate, so let us turn now to the subadditivity. With all the upper limits below understood to be taken as \( y \rightarrow x \) and \( t \downarrow 0 \), we calculate: \[ {f}^{ \circ }\left( {x;v + w}\right) = \lim \sup \frac{f\left( {y + {tv} + {tw}}\right) - f\left( y\right) }{t} \] \[ \leq \limsup \frac{f\left( {y + {tv} + {tw}}\right) - f\left( {y + {tw}}\right) }{t} + \limsup \frac{f\left( {y + {tw}}\right) - f\left( y\right) }{t} \] (since the upper limit of a sum is bounded above by the sum of the upper limits). The first upper limit in this last expression is \( {f}^{ \circ }\left( {x;v}\right) \), since the term \( y + {tw} \) represents in essence just a dummy variable converging to \( x \) . We conclude \[ {f}^{ \circ }\left( {x;v + w}\right) \leq {f}^{ \circ }\left( {x;v}\right) + {f}^{ \circ }\left( {x;w}\right) . \] which establishes (a). Now let \( \left\{ {x}_{i}\right\} \) and \( \left\{ {v}_{i}\right\} \) be arbitrary sequences converging to \( x \) and \( v \), respectively. For each \( i \), by definition of the upper limit, there exist \( {y}_{i} \) in \( X \) and \( {t}_{i} > 0 \) such that \[ \begin{Vmatrix}{{y}_{i} - {x}_{i}}\end{Vmatrix} + {t}_{i} < \frac{1}{i} \] \[ {f}^{ \circ }\left( {{x}_{i};{v}_{i}}\right) - \frac{1}{i} \leq \frac{f\left( {{y}_{i} + {t}_{i}{v}_{i}}\right) - f\left( {y}_{i}\right) }{{t}_{i}} \] \[ = \frac{f\left( {{y}_{i} + {t}_{i}v}\right) - f\left( {y}_{i}\right) }{{t}_{i}} + \frac{f\left( {{y}_{i} + {t}_{i}{v}_{i}}\right) - f\left( {{y}_{i} + {t}_{i}v}\right) }{{t}_{i}}. \] Note that the last term is bounded in magnitude by \( K\begin{Vmatrix}{
1012_(GTM178)Nonsmooth Analysis and Control Theory
25
rm \( y + {tw} \) represents in essence just a dummy variable converging to \( x \) . We conclude \[ {f}^{ \circ }\left( {x;v + w}\right) \leq {f}^{ \circ }\left( {x;v}\right) + {f}^{ \circ }\left( {x;w}\right) . \] which establishes (a). Now let \( \left\{ {x}_{i}\right\} \) and \( \left\{ {v}_{i}\right\} \) be arbitrary sequences converging to \( x \) and \( v \), respectively. For each \( i \), by definition of the upper limit, there exist \( {y}_{i} \) in \( X \) and \( {t}_{i} > 0 \) such that \[ \begin{Vmatrix}{{y}_{i} - {x}_{i}}\end{Vmatrix} + {t}_{i} < \frac{1}{i} \] \[ {f}^{ \circ }\left( {{x}_{i};{v}_{i}}\right) - \frac{1}{i} \leq \frac{f\left( {{y}_{i} + {t}_{i}{v}_{i}}\right) - f\left( {y}_{i}\right) }{{t}_{i}} \] \[ = \frac{f\left( {{y}_{i} + {t}_{i}v}\right) - f\left( {y}_{i}\right) }{{t}_{i}} + \frac{f\left( {{y}_{i} + {t}_{i}{v}_{i}}\right) - f\left( {{y}_{i} + {t}_{i}v}\right) }{{t}_{i}}. \] Note that the last term is bounded in magnitude by \( K\begin{Vmatrix}{{v}_{i} - v}\end{Vmatrix} \) (in view of the Lipschitz condition). Upon taking upper limits (as \( i \rightarrow \infty \) ), we derive \[ \mathop{\limsup }\limits_{{i \rightarrow \infty }}{f}^{ \circ }\left( {{x}_{i};{v}_{i}}\right) \leq {f}^{ \circ }\left( {x;v}\right) \] which establishes the upper semicontinuity. Finally, let any \( v \) and \( w \) in \( X \) be given. We have \[ f\left( {y + {tv}}\right) - f\left( y\right) \leq f\left( {y + {tw}}\right) - f\left( y\right) + K\parallel v - w\parallel t \] for \( y \) near \( x, t \) near 0 . Dividing by \( t \) and taking upper limits as \( y \rightarrow x \) , \( t \downarrow 0 \), gives \[ {f}^{ \circ }\left( {x;v}\right) \leq {f}^{ \circ }\left( {x;w}\right) + K\parallel v - w\parallel . \] Since this also holds with \( v \) and \( w \) switched,(b) follows. To prove (c), we calculate: \[ {f}^{ \circ }\left( {x; - v}\right) \mathrel{\text{:=}} \mathop{\limsup }\limits_{\substack{{{x}^{\prime } \rightarrow x} \\ {t \downarrow 0} }}\frac{f\left( {{x}^{\prime } - {tv}}\right) - f\left( {x}^{\prime }\right) }{t} \] \[ = \mathop{\limsup }\limits_{\substack{{u \rightarrow x} \\ {t \downarrow 0} }}\frac{\left( {-f}\right) \left( {u + {tv}}\right) - \left( {-f}\right) \left( u\right) }{t},\text{ where }u \mathrel{\text{:=}} {x}^{\prime } - {tv} \] \[ = {\left( -f\right) }^{ \circ }\left( {x;v}\right) \] as stated. 1.2. Exercise. Let \( f \) and \( g \) be Lipschitz near \( x \) . Prove that for any \( v \in X \) \[ {\left( f + g\right) }^{ \circ }\left( {x;v}\right) \leq {f}^{ \circ }\left( {x;v}\right) + {g}^{ \circ }\left( {x;v}\right) . \] A function such as \( v \mapsto {f}^{ \circ }\left( {x;v}\right) \) which is positively homogeneous and subadditive on \( X \) is the support function of a uniquely determined closed convex set in \( {X}^{ * } \) (the dual space of continuous linear functionals on \( X \) ). Some terminology is in order. Given a nonempty subset \( \sum \) of \( {X}^{ * } \), its support function is the function \( {H}_{\sum } : X \rightarrow ( - \infty ,\infty \rbrack \) defined as follows: \[ {H}_{\sum }\left( v\right) \mathrel{\text{:=}} \sup \{ \langle \zeta, v\rangle : \zeta \in \sum \} \] where we have used the familiar convention of denoting the value of the linear functional \( \zeta \) at \( v \) by \( \langle \zeta, v\rangle \) . We gather some useful facts about support functions in the next result. ## 1.3. Proposition. (a) Let \( \sum \) be a nonempty subset of \( {X}^{ * } \) . Then \( {H}_{\sum } \) is positively homogeneous, subadditive, and lower semicontinuous. (b) If \( \sum \) is convex and \( {w}^{ * } \) -closed, then a point \( \zeta \) in \( {X}^{ * } \) belongs to \( \sum \) iff we have \( {H}_{\sum }\left( v\right) \geq \langle \zeta, v\rangle \) for all \( v \) in \( X \) . (c) More generally, if \( \sum \) and \( \Lambda \) are two nonempty, convex, and \( {w}^{ * } \) -closed subsets of \( {X}^{ * } \), then \( \sum \supset \Lambda \) iff \( {H}_{\sum }\left( v\right) \geq {H}_{\Lambda }\left( v\right) \) for all \( v \) in \( X \) . (d) If \( p : X \rightarrow \mathbb{R} \) is positively homogeneous and subadditive and bounded on the unit ball, then there is a uniquely defined nonempty, convex, and \( {w}^{ * } \) -compact subset \( \sum \) of \( {X}^{ * } \) such that \( p = {H}_{\sum } \) . Proof. That \( {H}_{\sum } \) is positively homogeneous and subadditive follows immediately from its definition. As the upper envelope of continuous functions, \( {H}_{\sum } \) is automatically lower semicontinuous, whence (a). We turn now to (b), which is easily seen to amount to the following assertion: if \( \zeta \notin \sum \), then for some \( v \in X \) we have \( {H}_{\sum }\left( v\right) < \langle \zeta, v\rangle \) . This is proven by applying the Hahn-Banach Separation Theorem (see e.g., Rudin (1973)) to the topological vector space consisting of \( {X}^{ * } \) with its weak*-topology, bearing in mind that the dual of that space is identified with \( X \) . The proof of (c) is immediate in light of (b); there remains (d). Given \( p \), we set \[ \sum \mathrel{\text{:=}} \left\{ {\zeta \in {X}^{ * } : p\left( v\right) \geq \langle \zeta, v\rangle \forall v \in X}\right\} . \] Then \( \sum \) is seen to be convex as a consequence of the properties of \( p \), and \( {w}^{ * } \) -closed as the intersection of a family of \( {w}^{ * } \) -closed subsets. If \( K \) is a bound for \( p \) on \( B\left( {0;1}\right) \), then we have \( \langle \zeta, v\rangle \leq K \) for all \( v \in B\left( {0;1}\right) \), for any element \( \zeta \) of \( \sum \) . It follows that \( \sum \) is bounded, and hence \( {w}^{ * } \) -compact by Alaoglu’s Theorem. Clearly we have \( p \geq {H}_{\sum } \) ; let us prove equality. Let \( v \in X \) be given. Then, by a standard form of the Hahn-Banach Theorem (Rudin (1973, Theorem 3.2)), there exists \( \zeta \in {X}^{ * } \) such that \( \langle \zeta, w\rangle \leq p\left( w\right) \) \( \forall w \in X \), with \( \langle \zeta, v\rangle = p\left( v\right) \) . Then \( \zeta \in \sum \), so that \( {H}_{\sum }\left( v\right) = p\left( v\right) \) as required. Finally, the uniqueness of \( \sum \) follows from (c). Returning now to our function \( f \), and taking for the function \( p \) of the proposition the function \( {f}^{ \circ }\left( {x; \cdot }\right) \), we define the generalized gradient of \( f \) at \( x \), denoted \( \partial f\left( x\right) \), to be the (nonempty) \( {w}^{ * } \) -compact subset of \( {X}^{ * } \) whose support function is \( {f}^{ \circ }\left( {x; \cdot }\right) \) . Thus \( \zeta \in \partial f\left( x\right) \) iff \( {f}^{ \circ }\left( {x;v}\right) \geq \langle \zeta, v\rangle \) for all \( v \) in \( X \) . Since \( {f}^{ \circ }\left( {x; \cdot }\right) \) does not depend on which one of two equivalent norms on \( X \) is chosen, it follows that \( \partial f\left( x\right) \) too is independent of the particular norm on \( X \) . Some immediate intuition about \( \partial f \) is available from the following exercise, where we see that the relationship between \( {f}^{ \circ } \) and \( \partial f \) generalizes the classical formula \( {f}^{\prime }\left( {x;v}\right) = \left\langle {{f}^{\prime }\left( x\right), v}\right\rangle \) for the directional derivative \( {f}^{\prime }\left( {x;v}\right) \) . 1.4. Exercise. (a) Let \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be \( {C}^{1} \) . Prove that \( {f}^{ \circ }\left( {x;v}\right) = \left\langle {{f}^{\prime }\left( x\right), v}\right\rangle \) and that \( \partial f\left( x\right) = \left\{ {{f}^{\prime }\left( x\right) }\right\} \) . (b) Let \( f : \mathbb{R} \rightarrow \mathbb{R} \) be given by \( f\left( x\right) = \max \{ 0, x\} \) . Prove that \( {f}^{ \circ }\left( {x;v}\right) = \max \{ 0, v\} \) . What is \( \partial f\left( 0\right) \) ? (c) Find \( {f}^{ \circ }\left( {0; \cdot }\right) \) and \( \partial f\left( 0\right) \) when \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is given by \( f\left( x\right) = \) \( \parallel x\parallel \) . (d) If \( f \) has a local minimum or maximum at \( x \), then \( 0 \in \partial f\left( x\right) \) . We proceed now to derive some of the basic properties of the generalized gradient. A multivalued function \( F \) is said to be upper semicontinuous at \( x \) if for all \( \varepsilon > 0 \) there exists \( \delta > 0 \) such that \[ \parallel x - y\parallel < \delta \Rightarrow F\left( y\right) \subset F\left( x\right) + {\varepsilon B}. \] We denote by \( \parallel \zeta {\parallel }_{ * } \) the norm in \( {X}^{ * } \) : \[ \parallel \zeta {\parallel }_{ * } \mathrel{\text{:=}} \sup \{ \langle \zeta, v\rangle : v \in X,\parallel v\parallel = 1\} \] and \( {B}_{ * } \) denotes the open unit ball in \( {X}^{ * } \) . 1.5. Proposition. Let \( f \) be Lipschitz of rank \( K \) near \( x \) . Then: (a) \( \partial f\left( x\right) \) is a nonempty, convex, weak*-compact subset of \( {X}^{ * } \), and \( \parallel \zeta {\parallel }_{ * } \leq K \) for every \( \zeta \in \partial f\left( x\right) . \) (b) For every \( v \) in \( X \) we have \[ {f}^{ \circ }\left( {x;v}\right) = \max \{ \langle \zeta, v\rangle : \zeta \in \partial f\left( x\right) \} . \] (c) \( \zeta \in \partial f\left( x\right) \) iff \( {f}^{ \circ }\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v \in X \) . (d) If \( \left\{ {x}_{i}\right\} \) and \( \left\{ {\zeta }_{i}\right\} \) are sequences in \( X \) and \( {X}^{ * } \) such that \( {\zeta }_{i} \in \partial f\left( {x}_{i}\right) \) for each \( i \), and if \( {x}_{i} \) converges to \( x \) and \( \zeta \) is a weak* cluster point of the sequence \( \left\{ {\zeta }_{i}\right\} \), then we have \( \zeta \in \partial f\left( x\right) \) . (e) If \( X \) is finite dimensional, then \( \partial f \) is upper semicontinuous at \( x \) . Proof. We have already no
1012_(GTM178)Nonsmooth Analysis and Control Theory
26
x\right) \) is a nonempty, convex, weak*-compact subset of \( {X}^{ * } \), and \( \parallel \zeta {\parallel }_{ * } \leq K \) for every \( \zeta \in \partial f\left( x\right) . \) (b) For every \( v \) in \( X \) we have \[ {f}^{ \circ }\left( {x;v}\right) = \max \{ \langle \zeta, v\rangle : \zeta \in \partial f\left( x\right) \} . \] (c) \( \zeta \in \partial f\left( x\right) \) iff \( {f}^{ \circ }\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v \in X \) . (d) If \( \left\{ {x}_{i}\right\} \) and \( \left\{ {\zeta }_{i}\right\} \) are sequences in \( X \) and \( {X}^{ * } \) such that \( {\zeta }_{i} \in \partial f\left( {x}_{i}\right) \) for each \( i \), and if \( {x}_{i} \) converges to \( x \) and \( \zeta \) is a weak* cluster point of the sequence \( \left\{ {\zeta }_{i}\right\} \), then we have \( \zeta \in \partial f\left( x\right) \) . (e) If \( X \) is finite dimensional, then \( \partial f \) is upper semicontinuous at \( x \) . Proof. We have already noted that \( \partial f\left( x\right) \) is nonempty and \( {w}^{ * } \) -compact. Each \( \zeta \in \partial f\left( x\right) \) satisfies \( \langle \zeta, v\rangle \leq {f}^{ \circ }\left( {x;v}\right) \leq K\parallel v\parallel \) for all \( v \) in \( X \), whence \( \parallel \zeta {\parallel }_{ * } \leq K \) . The assertions (b),(c) merely reiterate that \( {f}^{ \circ }\left( {x; \cdot }\right) \) is the support function of \( \partial f\left( x\right) \) . Let us prove the closure property (d). Fix \( v \in X \) . For each \( i \), we have \( {f}^{ \circ }\left( {{x}_{i};v}\right) \geq \left\langle {{\zeta }_{i}, v}\right\rangle \) (in view of (c)). The sequence \( \left\{ \left\langle {{\zeta }_{i}, v}\right\rangle \right\} \) is bounded in \( \mathbb{R} \), and contains terms that are arbitrarily near \( \langle \zeta, v\rangle \) . Let us extract a subsequence of \( \left\{ {\zeta }_{i}\right\} \) (without relabeling) such that \( \left\langle {{\zeta }_{i}, v}\right\rangle \rightarrow \langle \zeta, v\rangle \) . Then passing to the limit in the preceding inequality gives \( {f}^{ \circ }\left( {x;v}\right) \geq \langle \zeta, v\rangle \), since \( {f}^{ \circ } \) is upper semicontinuous in \( x \) (Proposition 1.1). Since \( v \) is arbitrary, it follows (from (c) again) that \( \zeta \in \partial f\left( x\right) \) . We turn now to (e). Let \( \varepsilon > 0 \) be given; then we wish to show that for all \( y \) sufficiently near \( x \), we have \[ \partial f\left( y\right) \subset \partial f\left( x\right) + \varepsilon \bar{B} \] If this is not the case, then there is a sequence \( {y}_{i} \) converging to \( x \) and points \( {\zeta }_{i} \in \partial f\left( {y}_{i}\right) \) such that \( {\zeta }_{i} \notin \partial f\left( x\right) + \varepsilon \bar{B} \) . We can therefore separate \( {\zeta }_{i} \) from the compact convex set in question: for some \( {v}_{i} \neq 0 \) we have \[ \left\langle {{\zeta }_{i},{v}_{i}}\right\rangle \geq \max \left\{ {\left\langle {\zeta ,{v}_{i}}\right\rangle : \zeta \in \partial f\left( x\right) + \varepsilon \bar{B}}\right\} \] \[ = {f}^{ \circ }\left( {x;{v}_{i}}\right) + \varepsilon \begin{Vmatrix}{v}_{i}\end{Vmatrix}. \] Because of positive homogeneity, we can take \( \begin{Vmatrix}{v}_{i}\end{Vmatrix} = 1 \) . Note that the sequence \( \left\{ {\zeta }_{i}\right\} \) is bounded. Since we are in finite dimensions, we can extract convergent subsequences from \( \left\{ {\zeta }_{i}\right\} \) and \( \left\{ {v}_{i}\right\} \) (we do not relabel): \( {\zeta }_{i} \rightarrow \zeta \) , \( {v}_{i} \rightarrow v \), where \( \parallel v\parallel = 1 \) . The inequality above gives in the limit \( \langle \zeta, v\rangle \geq \) \( {f}^{ \circ }\left( {x;v}\right) + \varepsilon \), while invoking part (d) yields \( \zeta \in \partial f\left( x\right) \) . But then (c) is contradicted. This completes the proof. 1.6. Exercise. (a) Verify the upper semicontinuity of \( \partial f \) at 0 for each of the functions in Exercise 1.4(b, c). (b) Let \( {\zeta }_{i} \in \partial f\left( {x}_{i}\right) + {\varepsilon }_{i}{B}_{ * } \), where \( {x}_{i} \rightarrow x \) and \( {\varepsilon }_{i} \downarrow 0 \) . Let \( \zeta \) be a weak*-cluster point of \( \left\{ {\zeta }_{i}\right\} \) . Prove that \( \zeta \in \partial f\left( x\right) \) . ## 2 Basic Calculus We will derive an assortment of formulas that facilitate the calculation of \( \partial f \) when \( f \) is synthesized from simpler functionals through linear combinations, maximization, composition, and so on. We always assume that the given functions are Lipschitz near the point of interest; as we will see, this property has the useful feature of being preserved under the operations in question. 2.1. Proposition. For any scalar \( \lambda \), we have \( \partial \left( {\lambda f}\right) \left( x\right) = \lambda \partial f\left( x\right) \) . Proof. Note that \( {\lambda f} \) is Lipschitz near \( x \), of rank \( \left| \lambda \right| K \) . When \( \lambda \) is nonnegative, \( {\left( \lambda f\right) }^{ \circ } = \lambda {f}^{ \circ } \), and the result follows immediately. To complete the proof, it suffices to consider now the case \( \lambda = - 1 \) . An element \( \zeta \) of \( {X}^{ * } \) belongs to \( \partial \left( {-f}\right) \left( x\right) \) iff \( {\left( -f\right) }^{ \circ }\left( {x;v}\right) \geq \langle \zeta, v\rangle \) for all \( v \) . By Proposition 1.1(c), this is equivalent to: \( {f}^{ \circ }\left( {x; - v}\right) \geq \langle \zeta, v\rangle \) for all \( v \), which is equivalent to \( - \zeta \) belonging to \( \partial f\left( x\right) \) (by Proposition 1.3(c).) We now examine the generalized gradient of the sum of two functions \( f \) and \( g \), each of which is Lipschitz near \( x \) . It is easy to see that \( f + g \) is also Lipschitz near \( x \), and we would like to relate \( \partial \left( {f + g}\right) \left( x\right) \) to \( \partial f\left( x\right) + \partial g\left( x\right) \) . We will now do so, and introduce a technique that will be used many times: that of proving an inclusion between closed convex sets by proving an equivalent inequality between support functions. The support function of \( \partial \left( {f + g}\right) \left( x\right) \), evaluated at \( v \), is \( {\left( f + g\right) }^{ \circ }\left( {x;v}\right) \) (by definition!), while that of \( \partial f\left( x\right) + \partial g\left( x\right) \) is \( {f}^{ \circ }\left( {x;v}\right) + {g}^{ \circ }\left( {x;v}\right) \) (the support function of a sum of sets is the sum of the support functions). Since the sum of two \( {w}^{ * } \) -compact sets is \( {w}^{ * } \) -compact (addition is \( {w}^{ * } \) -continuous on \( {X}^{ * } \times {X}^{ * } \) ), it follows that the general inequality \[ {\left( f + g\right) }^{ \circ }\left( {x;v}\right) \leq {f}^{ \circ }\left( {x;v}\right) + {g}^{ \circ }\left( {x;v}\right) \] noted in Exercise 1.2 is equivalent to the inclusion \[ \partial \left( {f + g}\right) \left( x\right) \subset \partial f\left( x\right) + \partial g\left( x\right) \] as observed in Proposition 1.3(c). The extension of this inclusion (a sum rule) to finite linear combinations is immediate. 2.2. Proposition. Let \( {f}_{i}\left( {i = 1,2,\ldots, n}\right) \), be Lipschitz near \( x \), and let \( {\lambda }_{i} \) \( \left( {i = 1,2,\ldots, n}\right) \) be scalars. Then \( f \mathrel{\text{:=}} \mathop{\sum }\limits_{{i = 1}}^{n}{\lambda }_{i}{f}_{i} \) is Lipschitz near \( x \), and we have \[ \partial \left( {\mathop{\sum }\limits_{{i = 1}}^{n}{\lambda }_{i}{f}_{i}}\right) \left( x\right) \subset \mathop{\sum }\limits_{{i = 1}}^{n}{\lambda }_{i}\partial {f}_{i}\left( x\right) \] 2.3. Exercise. Prove Proposition 2.2, and give an example with \( X = \mathbb{R} \) and \( n = 2 \) for which the inclusion is strict. 2.4. Theorem (Lebourg’s Mean Value Theorem). Let \( x \) and \( y \) belong to \( X \), and suppose that \( f \) is Lipschitz on an open set containing the line segment \( \left\lbrack {x, y}\right\rbrack \) . Then there exists a point \( u \) in \( \left( {x, y}\right) \) such that \[ f\left( y\right) - f\left( x\right) \in \langle \partial f\left( u\right), y - x\rangle . \] Proof. We will need the following special chain rule for the proof. We denote by \( {x}_{t} \) the point \( x + t\left( {y - x}\right) \) . Lemma. The function \( g : \left\lbrack {0,1}\right\rbrack \rightarrow R \) defined by \( g\left( t\right) = f\left( {x}_{t}\right) \) is Lipschitz on \( \left( {0,1}\right) \), and we have \[ \partial g\left( t\right) \subset \left\langle {\partial f\left( {x}_{t}\right), y - x}\right\rangle . \] Proof of the Lemma. The fact that \( g \) is Lipschitz is plain. The two closed convex sets appearing in the equation are in fact intervals in \( R \), so it suffices to prove that for \( v = \pm 1 \), we have \[ \max \{ \partial g\left( t\right) v\} \leq \max \left\{ {\left\langle {\partial f\left( {x}_{t}\right), y - x}\right\rangle v}\right\} \] Now the left-hand side is just \( {g}^{ \circ }\left( {t;v}\right) \) ; that is, \[ \mathop{\limsup }\limits_{\substack{{s \rightarrow t} \\ {\lambda \downarrow 0} }}\frac{g\left( {s + {\lambda v}}\right) - g\left( s\right) }{\lambda } \] \[ = \mathop{\limsup }\limits_{\substack{{s \rightarrow t} \\ {\lambda \downarrow 0} }}\frac{f\left( {x + \left\lbrack {s + {\lambda v}}\right\rbrack \left( {y - x}\right) }\right) - f\left( {x + s\left( {y - x}\right) }\right) }{\lambda } \] \[ \leq \mathop{\limsup }\limits_{\substack{{{y}^{\prime } \rightarrow {x}_{t}} \\ {\lambda \downarrow 0} }}\frac{f\left( {{y}^{\prime } + {\lambda v}\left( {y - x}\right) }\right) - f\left( {y}^{\prime }\right) }{\lambda } \] \[ = {f}^{ \circ }\left( {{x}_{t};v\left( {y - x}\right) }\right) \] \[ = \max \left\langle {\partial f\left( {x}_{t}\right), v\left( {y - x}\right) }\right\rangle \] which comp
1012_(GTM178)Nonsmooth Analysis and Control Theory
27
t) v\} \leq \max \left\{ {\left\langle {\partial f\left( {x}_{t}\right), y - x}\right\rangle v}\right\} \] Now the left-hand side is just \( {g}^{ \circ }\left( {t;v}\right) \) ; that is, \[ \mathop{\limsup }\limits_{\substack{{s \rightarrow t} \\ {\lambda \downarrow 0} }}\frac{g\left( {s + {\lambda v}}\right) - g\left( s\right) }{\lambda } \] \[ = \mathop{\limsup }\limits_{\substack{{s \rightarrow t} \\ {\lambda \downarrow 0} }}\frac{f\left( {x + \left\lbrack {s + {\lambda v}}\right\rbrack \left( {y - x}\right) }\right) - f\left( {x + s\left( {y - x}\right) }\right) }{\lambda } \] \[ \leq \mathop{\limsup }\limits_{\substack{{{y}^{\prime } \rightarrow {x}_{t}} \\ {\lambda \downarrow 0} }}\frac{f\left( {{y}^{\prime } + {\lambda v}\left( {y - x}\right) }\right) - f\left( {y}^{\prime }\right) }{\lambda } \] \[ = {f}^{ \circ }\left( {{x}_{t};v\left( {y - x}\right) }\right) \] \[ = \max \left\langle {\partial f\left( {x}_{t}\right), v\left( {y - x}\right) }\right\rangle \] which completes the proof of the lemma. Now to the proof of the theorem. Consider the function \( \theta \) on \( \left\lbrack {0,1}\right\rbrack \) defined by \[ \theta \left( t\right) = f\left( {x}_{t}\right) + t\left\lbrack {f\left( x\right) - f\left( y\right) }\right\rbrack . \] Note that \( \theta \left( 0\right) = \theta \left( 1\right) = f\left( x\right) \), so that there is a point \( t \) in \( \left( {0,1}\right) \) at which \( \theta \) attains a local minimum or maximum (by continuity). By Exercise 1.4(d) we have \( 0 \in \partial \theta \left( t\right) \) . We may calculate \( \partial \theta \left( t\right) \) by appealing to Propositions 2.1 and 2.2 , and the lemma. We deduce \[ 0 \in f\left( x\right) - f\left( y\right) + \left\langle {\partial f\left( {x}_{t}\right), y - x}\right\rangle \] which is the assertion of the theorem (take \( u = {x}_{t} \) ). 2.5. Theorem (The Chain Rule). Let \( F : X \rightarrow {\mathbb{R}}^{n} \) be Lipschitz near \( x \), and let \( g : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be Lipschitz near \( F\left( x\right) \) . Then the function \( f\left( {x}^{\prime }\right) \mathrel{\text{:=}} \) \( g\left( {F\left( {x}^{\prime }\right) }\right) \) is Lipschitz near \( x \), and we have \[ \partial f\left( x\right) \subset {\overline{\operatorname{co}}}^{ * }\{ \partial \langle \gamma, F\left( \cdot \right) \rangle \left( x\right) : \gamma \in \partial g\left( {F\left( x\right) }\right) \} \] where \( {\overline{\mathrm{{co}}}}^{ * } \) signifies the \( {w}^{ * } \) -closed convex hull. Proof. The fact that \( f \) is Lipschitz is left to the reader to show. Again it is an inclusion between two convex weak*-compact sets that is at issue; the corresponding support function inequality amounts to the statement that for given \( v \), there exists \( \gamma \) in \( \partial g\left( {F\left( x\right) }\right) \), and \( \zeta \) in the generalized gradient at \( x \) of the function \( {x}^{\prime } \mapsto \left\langle {\gamma, F\left( {x}^{\prime }\right) }\right\rangle \), such that \( {f}^{ \circ }\left( {x;v}\right) \leq \langle \zeta, v\rangle \) . We will prove the theorem by producing such a pair \( \gamma \) and \( \zeta \) . To begin, we give ourselves sequences \( {y}_{i} \rightarrow x \) and \( {t}_{i} \downarrow 0 \) realizing the lim sup in the definition of \( {f}^{ \circ }\left( {x;v}\right) \) ; i.e., such that \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{f\left( {{y}_{i} + {t}_{i}v}\right) - f\left( {y}_{i}\right) }{{t}_{i}} = {f}^{ \circ }\left( {x;v}\right) . \] Applying the Mean Value Theorem 2.4 gives, for each \( i \), an element \( {\gamma }_{i} \in \) \( \partial g\left( {z}_{i}\right) \) such that \[ \frac{f\left( {{y}_{i} + {t}_{i}v}\right) - f\left( {y}_{i}\right) }{{t}_{i}} = \frac{g\left( {F\left( {{y}_{i} + {t}_{i}v}\right) }\right) - g\left( {F\left( {y}_{i}\right) }\right) }{{t}_{i}} \] \[ = \left\langle {{\gamma }_{i},\frac{F\left( {{y}_{i} + {t}_{i}v}\right) - F\left( {y}_{i}\right) }{{t}_{i}}}\right\rangle \] where \( {z}_{i} \) lies on the line segment joining \( F\left( {y}_{i}\right) \) and \( F\left( {{y}_{i} + {t}_{i}v}\right) \) . It follows that \( {z}_{i} \rightarrow F\left( x\right) \), and that for a suitable subsequence we have \( {\gamma }_{i} \rightarrow \gamma \in \partial g\left( {F\left( x\right) }\right) \) (we eschew relabeling). This is the required \( \gamma \) ; we turn now to exhibiting \( \zeta \) . By the Mean Value Theorem again, there exists \( {\zeta }_{i} \in \partial \langle \gamma, F\left( \cdot \right) \rangle \left( {w}_{i}\right) \) such that \[ \left\langle {\gamma ,\frac{F\left( {{y}_{i} + {t}_{i}v}\right) - F\left( {y}_{i}\right) }{{t}_{i}}}\right\rangle = \left\langle {{\zeta }_{i}, v}\right\rangle \] where \( {w}_{i} \) is on the line segment joining \( {y}_{i} \) and \( {y}_{i} + {t}_{i}v \) . It follows that \( {w}_{i} \rightarrow \) \( x \), that the sequence \( \left\{ {\zeta }_{i}\right\} \) is bounded in \( {X}^{ * } \), and that \( \left\{ \left\langle {{\zeta }_{i}, v}\right\rangle \right\} \) is bounded in \( \mathbb{R} \) . We may pass again to a subsequence to arrange for \( \left\langle {{\zeta }_{i}, v}\right\rangle \) to converge to some limit; having done so, let \( \zeta \) be a weak*-cluster point of \( \left\{ {\zeta }_{i}\right\} \) . Then \( \left\langle {{\zeta }_{i}, v}\right\rangle \rightarrow \langle \zeta, v\rangle \) necessarily, and \( \zeta \in \partial \langle \gamma, F\left( \cdot \right) \rangle \left( x\right) \) (Proposition 1.5(d)). Combining the above, we arrive at \[ \frac{f\left( {{y}_{i} + {t}_{i}v}\right) - f\left( {y}_{i}\right) }{{t}_{i}} = \left\langle {\left( {{\gamma }_{i} - \gamma }\right) + \gamma ,\frac{F\left( {{y}_{i} + {t}_{i}v}\right) - F\left( {y}_{i}\right) }{{t}_{i}}}\right\rangle \] \[ = \left\langle {{\gamma }_{i} - \gamma ,\frac{F\left( {{y}_{i} + {t}_{i}v}\right) - F\left( {y}_{i}\right) }{{t}_{i}}}\right\rangle + \left\langle {{\zeta }_{i}, v}\right\rangle . \] Now the term \( \left\lbrack {F\left( {{y}_{i} + {t}_{i}v}\right) - F\left( {y}_{i}\right) }\right\rbrack /{t}_{i} \) is bounded because \( F \) is Lipschitz, and we know \( {\gamma }_{i} \rightarrow \gamma \) . Therefore passing to the limit yields \[ {f}^{ \circ }\left( {x;v}\right) = \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{f\left( {{y}_{i} + {t}_{i}v}\right) - f\left( {y}_{i}\right) }{{t}_{i}} = \langle \zeta, v\rangle \] which confirms that \( \zeta \) has the required properties. The following exercise illustrates the use of the Chain Rule in a special case; invoke Theorem 2.5 to deduce it. 2.6. Exercise. Let \( f \) and \( g \) be Lipschitz near \( x \) . Then the product function \( {fg} \) is Lipschitz near \( x \) and we have \[ \partial \left( {fg}\right) \left( x\right) \subset \partial \left( {f\left( x\right) g\left( \cdot \right) + g\left( x\right) f\left( \cdot \right) }\right) \left( x\right) \subset f\left( x\right) \partial g\left( x\right) + g\left( x\right) \partial f\left( x\right) . \] ## 3 Relation to Derivatives We remind the reader that some basic definitions and facts about classical differentiability were recalled in Chapter 1, §2. (These carry over to the present Banach space setting when the \( \langle \cdot , \cdot \rangle \) is given the duality pairing interpretation.) 3.1. Proposition. Let \( f \) be Lipschitz near \( x \) . (a) If \( f \) admits a Gâteaux derivative \( {f}_{G}^{\prime }\left( x\right) \) at \( x \), then \( {f}_{G}^{\prime }\left( x\right) \in \partial f\left( x\right) \) . (b) If \( f \) is continuously differentiable at \( x \), then \( \partial f\left( x\right) = \left\{ {{f}^{\prime }\left( x\right) }\right\} \) . Proof. By definition we have the following relation between \( {f}_{G}^{\prime }\left( x\right) \) and the one-sided directional derivatives: \[ {f}^{\prime }\left( {x;v}\right) = \left\langle {{f}_{G}^{\prime }\left( x\right), v}\right\rangle \forall v \in {\mathbb{R}}^{n}. \] But clearly, \( {f}^{\prime }\left( {x;v}\right) \leq {f}^{ \circ }\left( {x;v}\right) \) . That \( {f}_{G}^{\prime }\left( x\right) \) belongs to \( \partial f\left( x\right) \) now follows from Proposition 1.5(c). Now suppose that \( f \) is \( {C}^{1} \) in a neighborhood of \( x \), and fix \( v \in X \) . For \( y \) near \( x \) and \( t > 0 \) near 0, we have \[ \frac{f\left( {y + {tv}}\right) - f\left( y\right) }{t} = \left\langle {{f}^{\prime }\left( z\right), v}\right\rangle \] for some \( z \in \left( {y, y + {tv}}\right) \), by the classical Mean Value Theorem. As \( y \rightarrow x \) and \( t \downarrow 0 \), the point \( z \) converges to \( x \), and because \( {f}^{\prime }\left( \cdot \right) \) is continuous (as a map between the Banach spaces \( X \) and \( \left. {X}^{ * }\right) \), we derive \( {f}^{ \circ }\left( {x;v}\right) \leq \left\langle {{f}^{\prime }\left( x\right), v}\right\rangle \) . It follows now from Proposition 1.5(c) that \( \langle \zeta, v\rangle \leq \left\langle {{f}^{\prime }\left( x\right), v}\right\rangle \) whenever \( \zeta \in \partial f\left( x\right) \) . Since \( v \) is arbitrary, we conclude that \( \partial f\left( x\right) \) is the singleton \( \left\{ {{f}^{\prime }\left( x\right) }\right\} \) . Remark. In the end-of-chapter problems we will see that \( \partial f\left( x\right) \) reduces to a singleton precisely when \( f \) is "strictly differentiable" at \( x \), a notion intermediate between Gâteaux and continuous differentiability. 3.2. Theorem. Let \( F : X \rightarrow Y \) be continuously differentiable near \( x \), with \( Y \) a Banach space, and let \( g : Y \rightarrow \mathbb{R} \) be Lipschitz near \( F\left( x\right) \) . Then \( f \mathrel{\text{:=}} \) \( g \circ F \) is Lipschitz near \( x \), and we have \[ \partial f\left( x\right) \subset {\left( {F}^{\prime }\left( x\right) \right) }^{ * }\partial g\left( {F\left( x\right) }\righ
1012_(GTM178)Nonsmooth Analysis and Control Theory
28
from Proposition 1.5(c) that \( \langle \zeta, v\rangle \leq \left\langle {{f}^{\prime }\left( x\right), v}\right\rangle \) whenever \( \zeta \in \partial f\left( x\right) \) . Since \( v \) is arbitrary, we conclude that \( \partial f\left( x\right) \) is the singleton \( \left\{ {{f}^{\prime }\left( x\right) }\right\} \) . Remark. In the end-of-chapter problems we will see that \( \partial f\left( x\right) \) reduces to a singleton precisely when \( f \) is "strictly differentiable" at \( x \), a notion intermediate between Gâteaux and continuous differentiability. 3.2. Theorem. Let \( F : X \rightarrow Y \) be continuously differentiable near \( x \), with \( Y \) a Banach space, and let \( g : Y \rightarrow \mathbb{R} \) be Lipschitz near \( F\left( x\right) \) . Then \( f \mathrel{\text{:=}} \) \( g \circ F \) is Lipschitz near \( x \), and we have \[ \partial f\left( x\right) \subset {\left( {F}^{\prime }\left( x\right) \right) }^{ * }\partial g\left( {F\left( x\right) }\right) , \] where \( {}^{ * } \) denotes the adjoint. If \( {F}^{\prime }\left( x\right) : X \rightarrow Y \) is onto, then equality holds. Proof. The fact that \( f \) is Lipschitz near \( x \) is straightforward. In terms of support functions, we must prove that given any \( v \), then we have some element \( \zeta \) of \( \partial g\left( {F\left( x\right) }\right) \) such that \[ {f}^{ \circ }\left( {x;v}\right) \leq \left\langle {v,{F}^{\prime }{\left( x\right) }^{ * }\zeta }\right\rangle = \left\langle {\zeta ,{F}^{\prime }\left( x\right) v}\right\rangle . \] For any \( {y}_{i} \) near \( x \) and \( {t}_{i} > 0 \) near 0, the difference quotient \( \left\lbrack {f\left( {{y}_{i} + {t}_{i}v}\right) - }\right. \) \( \left. {f\left( {y}_{i}\right) }\right\rbrack /{t}_{i} \) can be expressed in the form \[ \left\langle {{\zeta }_{i},\frac{F\left( {{y}_{i} + {t}_{i}v}\right) - F\left( {y}_{i}\right) }{{t}_{i}}}\right\rangle \] for some \( {\zeta }_{i} \in \partial g\left( {z}_{i}\right) \), where \( {z}_{i} \) lies in the segment \( \left\lbrack {F\left( {y}_{i}\right), F\left( {{y}_{i} + {t}_{i}v}\right) }\right\rbrack \) (we have invoked the Mean Value Theorem 2.4, of course). Much as in the proof of Theorem 2.5, we extract a subsequence from \( \left\{ {\zeta }_{i}\right\} \) such that \( \left\langle {{\zeta }_{i}, v}\right\rangle \) converges, and then let \( \zeta \) be a cluster point of \( {\zeta }_{i} \) . The required inequality follows from the fact that \( \left\lbrack {F\left( {{y}_{i} + {t}_{i}v}\right) - F\left( {y}_{i}\right) }\right\rbrack /{t}_{i} \) converges (strongly) to \( {F}^{\prime }\left( x\right) v \) . Now suppose that \( {F}^{\prime }\left( x\right) \) is onto. It follows that \( F \) maps every neighborhood of \( x \) to a neighborhood of \( F\left( x\right) \), by a classical theorem due to Graves (we will prove this and other such surjectivity results in Chapter 3). This fact justifies the second equality below: \[ {g}^{ \circ }\left( {F\left( x\right) ;{F}^{\prime }\left( x\right) v}\right) = \mathop{\limsup }\limits_{\substack{{y \rightarrow F\left( x\right) } \\ {t \downarrow 0} }}\frac{g\left( {y + t{F}^{\prime }\left( x\right) v}\right) - g\left( y\right) }{t} \] \[ = \mathop{\limsup }\limits_{\substack{{{x}^{\prime } \rightarrow x} \\ {t \downarrow 0} }}\frac{g\left( {F\left( {x}^{\prime }\right) + t{F}^{\prime }\left( x\right) v}\right) - g\left( {F\left( {x}^{\prime }\right) }\right) }{t} \] \[ = \mathop{\limsup }\limits_{\substack{{{x}^{\prime } \rightarrow x} \\ {t \downarrow 0} }}\frac{g\left( {F\left( {{x}^{\prime } + {tv}}\right) }\right) - g\left( {F\left( {x}^{\prime }\right) }\right) }{t} \] (since \( \left\lbrack {F\left( {{x}^{\prime } + {tv}}\right) - F\left( {x}^{\prime }\right) - t{F}^{\prime }\left( x\right) v}\right\rbrack /t \) goes to zero as \( {x}^{\prime } \rightarrow x \) and \( t \downarrow 0 \) , and since \( g \) is Lipschitz locally) \[ = {f}^{ \circ }\left( {x;v}\right) \text{.} \] Since \( v \) is arbitrary, this implies equality between the two sets figuring in the statement of the theorem, as asserted. The fact that equality can hold in the estimate of Theorem 3.2 under a supplementary hypothesis raises as a more general question the possibility of being able to assert equalities in other results too, for example in Proposition 2.2 and Theorem 2.5, or perhaps in Theorem 3.2, but with a supplementary hypothesis on \( g \) rather than \( F \) . Of course if the data are all smooth then equality holds in all these results, for then they simply reduce to classical differentiation formulas. The interest would lie in identifying a class of functions, not necessarily smooth, giving rise to equalities even when nonsingleton sets are involved. We address this issue in the next section. ## 4 Convex and Regular Functions A real-valued function \( f \) defined on an open convex subset \( U \) of \( X \) is termed convex provided that for any two points \( x, y \in U \) we have \[ f\left( {{tx} + \left( {1 - t}\right) y}\right) \leq {tf}\left( x\right) + \left( {1 - t}\right) f\left( y\right) \forall t \in \left\lbrack {0,1}\right\rbrack . \] The proof of Proposition 1.7.1 goes through without change in a Banach space to give: 4.1. Proposition. If \( f \) is a convex function on \( U \) that is bounded above on a neighborhood of some point in \( U \), then for any \( x \) in \( U, f \) is Lipschitz near \( x \) . 4.2. Exercise. Let \( f : X \rightarrow \mathbb{R} \) be convex, and let \( \theta : \mathbb{R} \rightarrow \mathbb{R} \) be convex and nondecreasing. Prove that \( x \mapsto \theta \left( {f\left( x\right) }\right) \) is convex. 4.3. Proposition. Let \( f \) be convex on \( U \) and Lipschitz near \( x \in U \) . Then the directional derivatives \( {f}^{\prime }\left( {x;v}\right) \) exist, and we have \( {f}^{\prime }\left( {x;v}\right) = {f}^{ \circ }\left( {x;v}\right) \) . A vector \( \zeta \) belongs to \( \partial f\left( x\right) \) iff \[ f\left( y\right) - f\left( x\right) \geq \langle \zeta, y - x\rangle \forall y \in U. \] Proof. It follows directly from the definition of convex function that for small \( t > 0 \), the function \[ t \mapsto \frac{f\left( {{x}^{\prime } + {tv}}\right) - f\left( {x}^{\prime }\right) }{t} \] is nondecreasing. This fact, together with the Lipschitz hypothesis, implies the existence and finiteness of the directional derivative for all \( {x}^{\prime } \) near \( x \) , for all \( v \) : \[ {f}^{\prime }\left( {{x}^{\prime };v}\right) = \mathop{\inf }\limits_{{t > 0}}\frac{f\left( {{x}^{\prime } + {tv}}\right) - f\left( {x}^{\prime }\right) }{t}. \] Now fix \( \delta > 0 \), and observe that \( {f}^{ \circ }\left( {x;v}\right) \) can be written as \[ {f}^{ \circ }\left( {x;v}\right) = \mathop{\lim }\limits_{{\varepsilon \downarrow 0}}\mathop{\sup }\limits_{{\begin{Vmatrix}{{x}^{\prime } - x}\end{Vmatrix} \leq {\varepsilon \delta }}}\mathop{\sup }\limits_{{0 < t < \varepsilon }}\frac{f\left( {{x}^{\prime } + {tv}}\right) - f\left( {x}^{\prime }\right) }{t}. \] The preceding remarks show that an alternative expression for \( {f}^{ \circ }\left( {x;v}\right) \) is \[ \mathop{\lim }\limits_{{\varepsilon \downarrow 0}}\mathop{\sup }\limits_{{\begin{Vmatrix}{{x}^{\prime } - x}\end{Vmatrix} \leq {\varepsilon \delta }}}\frac{f\left( {{x}^{\prime } + {\varepsilon v}}\right) - f\left( {x}^{\prime }\right) }{\varepsilon }. \] If \( K \) is a Lipschitz constant for \( f \) near \( x \), then for all \( {x}^{\prime } \) in \( B\left( {x;{\varepsilon \delta B}}\right) \), for all \( \varepsilon \) sufficiently small, we have \[ \left| {\frac{f\left( {{x}^{\prime } + {\varepsilon v}}\right) - f\left( {x}^{\prime }\right) }{\varepsilon } - \frac{f\left( {x + {\varepsilon v}}\right) - f\left( x\right) }{\varepsilon }}\right| \leq {2\delta K} \] so that \[ {f}^{ \circ }\left( {x;v}\right) \leq \mathop{\lim }\limits_{{\varepsilon \downarrow 0}}\left\{ {\frac{f\left( {x + {\varepsilon v}}\right) - f\left( x\right) }{\varepsilon } + {2\delta K}}\right\} = {f}^{\prime }\left( {x;v}\right) + {2\delta K}. \] Since \( \delta \) is arbitrary we deduce \( {f}^{ \circ }\left( {x;v}\right) \leq {f}^{\prime }\left( {x;v}\right) \), and hence equality, since \( {f}^{ \circ } \geq {f}^{\prime } \) inherently. Finally, we observe \[ \zeta \in \partial f\left( x\right) \Leftrightarrow {f}^{ \circ }\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v, \] \[ \Leftrightarrow {f}^{\prime }\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v \] \[ \Leftrightarrow \mathop{\inf }\limits_{{t > 0}}\frac{f\left( {x + {tv}}\right) - f\left( x\right) }{t} \geq \langle \zeta, v\rangle \forall v \] \[ \Leftrightarrow f\left( y\right) - f\left( x\right) \geq \langle \zeta, y - x\rangle \forall y \in U. \] It turns out that the property of having directional derivatives \( {f}^{\prime }\left( {x;v}\right) \) that coincide with \( {f}^{ \circ }\left( {x;v}\right) \) is precisely what is required to make our calculus rules "more exact." We give this property a name: the function \( f \) is regular at \( x \) provided that \( f \) is Lipschitz near \( x \) and admits directional derivatives \( {f}^{\prime }\left( {x;v}\right) \) at \( x \) for all \( v \), with \( {f}^{\prime }\left( {x;v}\right) = {f}^{ \circ }\left( {x;v}\right) \) . Evidently, functions which are continuously differentiable at \( x \) are regular at \( x \), since then \( {f}^{\prime }\left( {x;v}\right) = \) \( \left\langle {{f}^{\prime }\left( x\right), v}\right\rangle = {f}^{ \circ }\left( {x;v}\right) \) . Also, convex functions which are Lipschitz near \( x \) are regular there, by the preceding proposition. 4.4. Exercise. Give an example of a function which is neither \( {C}^{1} \) nor convex near \( x \), but which is regular at \( x \) . Let us now illustrate how regularity sharpens certain calculus rules, such as that for the sum of two function
1012_(GTM178)Nonsmooth Analysis and Control Theory
29
\left( {x;v}\right) \) is precisely what is required to make our calculus rules "more exact." We give this property a name: the function \( f \) is regular at \( x \) provided that \( f \) is Lipschitz near \( x \) and admits directional derivatives \( {f}^{\prime }\left( {x;v}\right) \) at \( x \) for all \( v \), with \( {f}^{\prime }\left( {x;v}\right) = {f}^{ \circ }\left( {x;v}\right) \) . Evidently, functions which are continuously differentiable at \( x \) are regular at \( x \), since then \( {f}^{\prime }\left( {x;v}\right) = \) \( \left\langle {{f}^{\prime }\left( x\right), v}\right\rangle = {f}^{ \circ }\left( {x;v}\right) \) . Also, convex functions which are Lipschitz near \( x \) are regular there, by the preceding proposition. 4.4. Exercise. Give an example of a function which is neither \( {C}^{1} \) nor convex near \( x \), but which is regular at \( x \) . Let us now illustrate how regularity sharpens certain calculus rules, such as that for the sum of two functions. If \( f \) and \( g \) are Lipschitz near \( x \), we know (Proposition 2.2) that \[ \partial \left( {f + g}\right) \left( x\right) \subset \partial f\left( x\right) + \partial g\left( x\right) . \] Suppose now that \( f \) and \( g \) are regular at \( x \) . Then we can argue as follows to get the opposite inclusion: for any \( v \) , \[ \max \{ \langle \zeta + \xi, v\rangle : \zeta \in \partial f\left( x\right) ,\xi \in \partial g\left( x\right) \} \] \[ = {f}^{ \circ }\left( {x;v}\right) + {g}^{ \circ }\left( {x;v}\right) \] \[ = {f}^{\prime }\left( {x;v}\right) + {g}^{\prime }\left( {x;v}\right) \] \[ = {\left( f + g\right) }^{\prime }\left( {x;v}\right) \leq {\left( f + g\right) }^{ \circ }\left( {x;v}\right) \] \[ = \max \{ \langle \zeta, v\rangle : \zeta \in \partial \left( {f + g}\right) \left( x\right) \} . \] This inequality between support functions is equivalent to the inclusion \[ \partial f\left( x\right) + \partial g\left( x\right) \subset \partial \left( {f + g}\right) \left( x\right) \] so that equality actually holds. A bonus consequence of this argument is the fact that \( {\left( f + g\right) }^{\prime }\left( {x; \cdot }\right) \) and \( {\left( f + g\right) }^{ \circ }\left( {x; \cdot }\right) \) coincide, so that \( f + g \) inherits regularity from \( f \) and \( g \) . In fact, it is clear that any (finite) nonnegative linear combination of regular functions is regular. The following theorem subsumes the case of a finite sum just discussed, and in its regularity conclusions is related to Exercise 4.2. The setting is that of the Chain Rule 2.5, of which this is a refinement. 4.5. Theorem. Let \( F : X \rightarrow {\mathbb{R}}^{n} \) be such that each component function \( {f}_{i} \) of \( F \) is regular at \( x \) . Let \( g : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be regular at \( F\left( x\right) \), and suppose that each \( \gamma \in \partial g\left( {F\left( x\right) }\right) \) has nonnegative components. Then the function \( f\left( {x}^{\prime }\right) \mathrel{\text{:=}} g\left( {F\left( {x}^{\prime }\right) }\right) \) is regular at \( x \), and we have \[ \partial f\left( x\right) = {\overline{\operatorname{co}}}^{ * }\{ \partial \langle \gamma, F\left( \cdot \right) \rangle \left( x\right) : \gamma \in \partial g\left( {F\left( x\right) }\right) \} . \] Proof. We ask the reader to check as a first step that \( f \) admits directional derivatives at \( x \) : \[ {f}^{\prime }\left( {x;v}\right) = {g}^{\prime }\left( {F\left( x\right) ;{F}^{\prime }\left( {x;v}\right) }\right) , \] where \( {F}^{\prime }\left( {x;v}\right) \) signifies the vector in \( {\mathbb{R}}^{n} \) whose \( i \) th component is \( {f}_{i}^{\prime }\left( {x;v}\right) \) . Now consider, for given \( v \in X \), the maximum of the inner product of \( v \) taken with elements from the right side of the equality asserted by the theorem. That maximum equals \[ \max \left\{ {\langle \gamma, F\left( \cdot \right) {\rangle }^{ \circ }\left( {x;v}\right) : \gamma \in \partial g\left( {F\left( x\right) }\right) }\right\} \] \[ = \max \left\{ {\langle \gamma, F\left( \cdot \right) {\rangle }^{\prime }\left( {x;v}\right) : \gamma \in \partial g\left( {F\left( x\right) }\right) }\right\} \] (for \( \langle \gamma, F\left( \cdot \right) \rangle \) is regular at \( x \), as a nonnegative linear combination of functions regular at \( x \), since \( \gamma \) is nonnegative) \[ = \max \left\{ {\left\langle {\gamma ,{F}^{\prime }\left( {x;v}\right) }\right\rangle : \gamma \in \partial g\left( {F\left( x\right) }\right) }\right\} \] \[ = {g}^{ \circ }\left( {F\left( x\right) ;{F}^{\prime }\left( {x;v}\right) }\right) \] \[ = {g}^{\prime }\left( {F\left( x\right) ;{F}^{\prime }\left( {x;v}\right) }\right) \;\left( {\text{since}g\text{ is regular at }F\left( x\right) }\right) \] \[ = {f}^{\prime }\left( {x;v}\right) \;\text{(as noted above)} \] \[ \leq {f}^{ \circ }\left( {x;v}\right) \text{.} \] But this last term is the support function of the left side, evaluated at \( v \) , implying the opposite inclusion to the one furnished by Theorem 2.5. It follows that the two sets coincide, and that \( {f}^{ \circ }\left( {x;v}\right) \) and \( {f}^{\prime }\left( {x;v}\right) \) agree. 4.6. Exercise. (a) Let \( g : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be nondecreasing in its first component, and let \( \zeta = \left( {{\zeta }_{1},{\zeta }_{2},\ldots ,{\zeta }_{n}}\right) \in \partial g\left( x\right) \) . Prove that \( {\zeta }_{1} \geq 0 \) . (b) Find conditions on \( f \) and \( g \) and their values at \( x \) under which equality holds in the estimate of Exercise 2.6. (c) Let \( {f}_{i}\left( {i = 1,2,\ldots, n}\right) \) be Lipschitz near \( x \), and set \( f\left( x\right) \mathrel{\text{:=}} \) \( \mathop{\max }\limits_{{1 \leq i \leq n}}{f}_{i}\left( x\right) \) . Prove that \( f \) is Lipschitz near \( x \), with \[ \partial f\left( x\right) \subset \operatorname{co}\left\{ {\mathop{\bigcup }\limits_{{i \in M\left( x\right) }}\partial {f}_{i}\left( x\right) }\right\} \] where \( M\left( x\right) = \left\{ {i \in \{ 1,2,\ldots, n\} : {f}_{i}\left( x\right) = f\left( x\right) }\right\} \) . Note that we must show that the set appearing on the right is \( {w}^{ * } \) -closed. If each \( {f}_{i} \) is regular at \( x \), prove that equality holds. (Hint. Problem 1.11.14.) Develop the analogous formula when max is replaced by min. (d) Let \( f : {\mathbb{R}}^{2} \rightarrow \mathbb{R} \) be given by \( f\left( {x, y}\right) \mathrel{\text{:=}} \max \left\{ {\left| {x - y}\right|, y - {x}^{2}}\right\} \) . Find \( \partial f\left( {0,0}\right) \) . ## 5 Tangents and Normals Let \( S \) be a nonempty closed subset of \( X \) . There is a globally Lipschitz function associated with \( S \) that completely characterizes it: its distance function \( {d}_{S}\left( \cdot \right) \) is given by \[ {d}_{S}\left( x\right) \mathrel{\text{:=}} \inf \{ \parallel x - s\parallel : s \in S\} . \] We can apply our Lipschitz calculus to \( {d}_{S}\left( \cdot \right) \) in order to define geometric constructs for \( S \) . In this light, a rather natural way to define a direction \( v \) tangent to \( S \) at \( x \in S \) is as follows: we require \( {d}_{S}^{ \circ }\left( {x;v}\right) \leq 0 \) . (That is, \( {d}_{S} \) should not increase in the \( v \) direction, as measured by the generalized directional derivative.) We remark that since \( {d}_{S}^{ \circ }\left( {x;v}\right) \geq 0 \) for all \( v \) (show), it is equivalent to require \( {d}_{S}^{ \circ }\left( {x;v}\right) = 0 \) . We proceed to adopt this definition: the tangent cone to \( S \) at \( x \), denoted \( {T}_{S}\left( x\right) \), is the set of all those \( v \in X \) satisfying \( {d}^{ \circ }\left( {x;v}\right) \leq 0 \) . 5.1. Exercise. Show that \( 0 \in {T}_{S}\left( x\right) \), and that \( {T}_{S}\left( x\right) \) is a cone. Prove that \( {T}_{S}\left( x\right) \) is closed and convex. It is occasionally useful to have the following alternate, direct, characterization of \( {T}_{S}\left( x\right) \) on hand, and reassuring to know that tangency does not depend on the choice of equivalent norms for \( X \) (as \( {d}_{S} \) does): 5.2. Proposition. An element \( v \) of \( X \) is tangent to \( S \) at \( x \) iff, for every sequence \( {x}_{i} \) in \( S \) converging to \( x \) and sequence \( {t}_{i} \) in \( \left( {0,\infty }\right) \) decreasing to 0, there exists a sequence \( {v}_{i} \) in \( X \) converging to \( v \) such that \( {x}_{i} + {t}_{i}{v}_{i} \in S \) for all \( i \) . Proof. Suppose first that \( v \in {T}_{S}\left( x\right) \), and that sequences \( {x}_{i} \rightarrow x \) (with \( \left. {{x}_{i} \in S}\right) ,{t}_{i} \downarrow 0 \) are given. We must produce the sequence \( {v}_{i} \) alluded to in the statement of the theorem. Since \( {d}_{S}^{ \circ }\left( {x;v}\right) = 0 \) by assumption, we have \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{{d}_{S}\left( {{x}_{i} + {t}_{i}v}\right) - {d}_{S}\left( {x}_{i}\right) }{{t}_{i}} = \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{{d}_{S}\left( {{x}_{i} + {t}_{i}v}\right) }{{t}_{i}} = 0. \] Let \( {s}_{i} \) be a point in \( S \) which satisfies \[ \begin{Vmatrix}{{x}_{i} + {t}_{i}v - {s}_{i}}\end{Vmatrix} \leq {d}_{S}\left( {{x}_{i} + {t}_{i}v}\right) + \frac{{t}_{i}}{i} \] and let us set \[ {v}_{i} = \frac{{s}_{i} - {x}_{i}}{{t}_{i}} \] Then \( \begin{Vmatrix}{v - {v}_{i}}\end{Vmatrix} \rightarrow 0 \) ; that is, \( {v}_{i} \) converges to \( v \) . Furthermore, \( {x}_{i} + {t}_{i}{v}_{i} = {s}_{i} \in \) \( S \), as required. Now for the converse. Let \( v \) have the stated property concerning sequences, and choose a sequence \( {y}_{i} \) converging to \( x \) and \( {t}_{i} \) decreasing to 0 such that \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{{d}_{S}\left( {{y}_{i} + {t}_{i}v}\right) - {d}_{S}\left(
1012_(GTM178)Nonsmooth Analysis and Control Theory
30
hop{\lim }\limits_{{i \rightarrow \infty }}\frac{{d}_{S}\left( {{x}_{i} + {t}_{i}v}\right) - {d}_{S}\left( {x}_{i}\right) }{{t}_{i}} = \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{{d}_{S}\left( {{x}_{i} + {t}_{i}v}\right) }{{t}_{i}} = 0. \] Let \( {s}_{i} \) be a point in \( S \) which satisfies \[ \begin{Vmatrix}{{x}_{i} + {t}_{i}v - {s}_{i}}\end{Vmatrix} \leq {d}_{S}\left( {{x}_{i} + {t}_{i}v}\right) + \frac{{t}_{i}}{i} \] and let us set \[ {v}_{i} = \frac{{s}_{i} - {x}_{i}}{{t}_{i}} \] Then \( \begin{Vmatrix}{v - {v}_{i}}\end{Vmatrix} \rightarrow 0 \) ; that is, \( {v}_{i} \) converges to \( v \) . Furthermore, \( {x}_{i} + {t}_{i}{v}_{i} = {s}_{i} \in \) \( S \), as required. Now for the converse. Let \( v \) have the stated property concerning sequences, and choose a sequence \( {y}_{i} \) converging to \( x \) and \( {t}_{i} \) decreasing to 0 such that \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{{d}_{S}\left( {{y}_{i} + {t}_{i}v}\right) - {d}_{S}\left( {y}_{i}\right) }{{t}_{i}} = {d}_{S}^{ \circ }\left( {x;v}\right) . \] Our purpose is to prove this quantity nonpositive, for then \( v \in {T}_{S}\left( x\right) \) by definition. Let \( {s}_{i} \) in \( S \) satisfy \[ \begin{Vmatrix}{{s}_{i} - {y}_{i}}\end{Vmatrix} \leq {d}_{S}\left( {y}_{i}\right) + \frac{{t}_{i}}{i} \] It follows that \( {s}_{i} \) converges to \( x \) . Thus there is a sequence \( {v}_{i} \) converging to \( v \) such that \( {s}_{i} + {t}_{i}{v}_{i} \in S \) . But then, since \( {d}_{S} \) is Lipschitz of rank 1, \[ {d}_{S}\left( {{y}_{i} + {t}_{i}v}\right) \leq {d}_{S}\left( {{s}_{i} + {t}_{i}{v}_{i}}\right) + \begin{Vmatrix}{{y}_{i} - {s}_{i}}\end{Vmatrix} + {t}_{i}\begin{Vmatrix}{v - {v}_{i}}\end{Vmatrix} \] \[ \leq {d}_{S}\left( {y}_{i}\right) + {t}_{i}\left( {\begin{Vmatrix}{v - {v}_{i}}\end{Vmatrix} + \frac{1}{i}}\right) . \] We deduce that the limit above is nonpositive, which completes the proof. 5.3. Exercise. Let \( X = {X}_{1} \times {X}_{2} \), where \( {X}_{1},{X}_{2} \) are Banach spaces, and let \( x = \left( {{x}_{1},{x}_{2}}\right) \in {S}_{1} \times {S}_{2} \), where \( {S}_{1},{S}_{2} \) are subsets of \( {X}_{1},{X}_{2} \) , respectively. Then \[ {T}_{{S}_{1} \times {S}_{2}}\left( x\right) = {T}_{{S}_{1}}\left( {x}_{1}\right) \times {T}_{{S}_{2}}\left( {x}_{2}\right) . \] ## The Normal Cone In the case of classical manifolds in \( {\mathbb{R}}^{n} \), the tangent space and the normal space are orthogonal to one another. When convex cones are involved, it is polarity that serves to obtain one from the other. We define the normal cone to \( S \) at \( x \), denoted \( {N}_{S}\left( x\right) \), as follows: \[ {N}_{S}\left( x\right) \mathrel{\text{:=}} {T}_{S}{\left( x\right) }^{ \circ } \mathrel{\text{:=}} \left\{ {\zeta \in {X}^{ * } : \langle \zeta, v\rangle \leq 0\forall v \in {T}_{S}\left( x\right) }\right\} . \] ## 5.4. Proposition. (a) \( {N}_{S}\left( x\right) \) is a \( {w}^{ * } \) -closed convex cone. (b) \( {N}_{S}\left( x\right) = {\operatorname{cl}}^{ * }\left\{ {\mathop{\bigcup }\limits_{{\lambda \geq 0}}\lambda \partial {d}_{S}\left( x\right) }\right\} \) . (c) \( {T}_{S}\left( x\right) \) is in turn the polar of \( {N}_{S}\left( x\right) \) ; that is, \[ {T}_{S}\left( x\right) = {N}_{S}{\left( x\right) }^{ \circ } = \left\{ {v \in X : \langle \zeta, v\rangle \leq 0\forall \zeta \in {N}_{S}\left( x\right) }\right\} . \] Proof. Property (a) is immediate. Let \( \zeta \in \partial {d}_{S}\left( x\right) \), and suppose that \( v \in \) \( {T}_{S}\left( x\right) \) . Since \( {d}_{S}^{ \circ }\left( {x;v}\right) \leq 0 \) by definition of \( {T}_{S}\left( x\right) \), and since \( {d}_{S}^{ \circ }\left( {x; \cdot }\right) \) is the support function of \( \partial {d}_{S}\left( x\right) \), we deduce \( \langle \zeta, v\rangle \leq 0 \) . This shows that \( \partial {d}_{S}\left( x\right) \) lies in \( {N}_{S}\left( x\right) \), which implies that the set \( \sum \) appearing on the right in (b) is contained in \( {N}_{S}\left( x\right) \) . To complete the proof of (b), let \( \zeta \) be a point in the complement of \( \sum \) . By the Separation Theorem, there exists \( v \in X \) such that \[ {H}_{\sum }\left( v\right) < \langle \zeta, v\rangle \] It follows that \( \langle \zeta, v\rangle > 0 \) and \( {H}_{\sum }\left( v\right) \leq 0 \), since \( \sum \) is a cone. Therefore \( \langle v,\theta \rangle \leq 0\forall \theta \in \partial {d}_{S}\left( x\right) \), whence \( {d}_{S}^{ \circ }\left( {x;v}\right) \leq 0 \) . We conclude that \( v \in {T}_{S}\left( x\right) \) . Since \( \langle \zeta, v\rangle > 0 \), it follows that \( \zeta \notin {N}_{S}\left( x\right) \) ; (b) is proven. We turn now to the proof of (c). Let \( v \in {T}_{S}\left( x\right) \) . Then \( \langle \zeta, v\rangle \leq 0\forall \zeta \in \) \( \partial {d}_{S}\left( x\right) \), which implies \( \langle \zeta, v\rangle \leq 0\forall \zeta \in {N}_{S}\left( x\right) \) in view of (b). Thus \( v \in \) \( {N}_{S}{\left( x\right) }^{ \circ } \) . Conversely, let \( v \in {N}_{S}{\left( x\right) }^{ \circ } \) . Then \( \langle \zeta, v\rangle \leq 0\forall \zeta \in \partial {d}_{S}\left( x\right) \), because of (b). But then \( {d}_{S}^{ \circ }\left( {x;v}\right) \leq 0 \) and \( v \in {T}_{S}\left( x\right) \) . We postpone to later the proof of the fact that \( {T}_{S} \) and \( {N}_{S} \) coincide with the classical tangent and normal spaces when \( S \) is a smooth manifold. Another special case of interest is the convex one, which we now examine. 5.5. Proposition. Let \( S \) be convex. Then \[ {T}_{S}\left( x\right) = \operatorname{cl}\{ \lambda \left( {s - x}\right) : \lambda \geq 0, s \in S\} \] and \[ {N}_{S}\left( x\right) = \left\{ {\zeta \in {X}^{ * } : \left\langle {\zeta ,{x}^{\prime } - x}\right\rangle \leq 0\forall {x}^{\prime } \in S}\right\} . \] Proof. The convexity of the set \( S \) readily implies that of the function \( {d}_{S}\left( \cdot \right) \) . It follows then from Proposition 4.3 that \( {d}_{S}^{\prime }\left( {x;v}\right) \) exists and coincides with \( {d}_{S}^{ \circ }\left( {x;v}\right) \) . Consequently, \( {T}_{S}\left( x\right) \) consists of those \( v \in X \) for which \[ \mathop{\lim }\limits_{{t \downarrow 0}}\frac{{d}_{S}\left( {x + {tv}}\right) }{t} = 0 \] This in turn is equivalent to the existence of \( s\left( t\right) \in S \) such that \[ \frac{\parallel x + {tv} - s\left( t\right) \parallel }{t} \rightarrow 0\text{ as }t \downarrow 0. \] Setting \( u\left( t\right) \mathrel{\text{:=}} \left( {x + {tv} - s\left( t\right) }\right) /t \), this can be expressed in the form \[ v = \left( \frac{1}{t}\right) \left( {s\left( t\right) - x}\right) + u\left( t\right) \] where \( u\left( t\right) \rightarrow 0 \) as \( t \downarrow 0 \) . This is equivalent to the characterization of \( {T}_{S}\left( x\right) \) given in the statement of the proposition. The expression for \( {N}_{S}\left( x\right) \) then follows immediately from this characterization, together with the fact that \( {N}_{S}\left( x\right) \) is the polar of \( {T}_{S}\left( x\right) \) . 5.6. Exercise. Determine the tangent and normal cones at the origin for each of the following subsets of \( {\mathbb{R}}^{2} \) : (a) \( {S}_{1} \mathrel{\text{:=}} \{ \left( {x, y}\right) : {xy} = 0\} \) , (b) \( {S}_{2} \mathrel{\text{:=}} \{ \left( {x, y}\right) : y \geq 2\left| x\right| \} \) , (c) \( {S}_{3} \mathrel{\text{:=}} \) the closure of the complement of the set \( {S}_{2} \) , (d) \( {S}_{4} \mathrel{\text{:=}} \{ \left( {x, y}\right) : y \leq \sqrt{\left| x\right| }\} \) , (e) \( {S}_{5} \mathrel{\text{:=}} \{ \left( {x, y}\right) : y = - \sqrt{\left| x\right| }\} \) , (f) \( {S}_{6} \mathrel{\text{:=}} \{ \left( {x, y}\right) : y \leq - \sqrt{\left| x\right| }\} \), and (g) \( {S}_{7} \mathrel{\text{:=}} {S}_{2} \cup \{ \left( {0, y}\right) : y \in \mathbb{R}\} \) . When \( S \) is the epigraph of a function, we would expect some relationship to exist between its tangent and normal cones on the one hand, and the generalized gradient of the function on the other. In fact, a complete duality exists, as we now see. 5.7. Theorem. Let \( f \) be Lipschitz near \( x \) . Then: (a) \( {T}_{\text{epi }f}\left( {x, f\left( x\right) }\right) = \operatorname{epi}{f}^{ \circ }\left( {x; \cdot }\right) \) ; and (b) \( \zeta \in \partial f\left( x\right) \Leftrightarrow \left( {\zeta , - 1}\right) \in {N}_{\text{epi }f}\left( {x, f\left( x\right) }\right) \) . Proof. Suppose first that \( \left( {v, r}\right) \) lies in \( {T}_{\text{epi }f}\left( {x, f\left( x\right) }\right) \) . Choose sequences \( {y}_{i} \rightarrow x,{t}_{i} \downarrow 0 \), such that \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{f\left( {{y}_{i} + {t}_{i}v}\right) - f\left( {y}_{i}\right) }{{t}_{i}} = {f}^{ \circ }\left( {x;v}\right) . \] Note that \( \left( {{y}_{i}, f\left( {y}_{i}\right) }\right) \) is a sequence in epi \( f \) converging to \( \left( {x, f\left( x\right) }\right) \) . Accordingly, by Proposition 5.2, there exists a sequence \( \left( {{v}_{i},{r}_{i}}\right) \) converging to \( \left( {v, r}\right) \) such that \( \left( {{y}_{i}, f\left( {y}_{i}\right) }\right) + {t}_{i}\left( {{v}_{i},{r}_{i}}\right) \in \operatorname{epi}f \) . Thus \[ f\left( {y}_{i}\right) + {t}_{i}{r}_{i} \geq f\left( {{y}_{i} + {t}_{i}{v}_{i}}\right) \] We rewrite this as \[ \frac{f\left( {{y}_{i} + {t}_{i}{v}_{i}}\right) - f\left( {y}_{i}\right) }{{t}_{i}} \leq {r}_{i} \] Taking limits, we obtain \( {f}^{ \circ }\left( {x;v}\right) \leq r \) as desired. We now show that for any \( v \), for any \( \delta \geq 0 \), the point \( \left( {v,{f}^{ \circ }\left( {x;v}\right) + \delta }\right) \) lies in \( {T}_{\text{epi }f}\left( {x, f\left( x\right) }\right) \) ; this will complete the proof of (a). Let \( \left( {{x}_{i},{r}_{i}}\right) \) be any sequence in epi \( f \)
1012_(GTM178)Nonsmooth Analysis and Control Theory
31
t) }\right) \) is a sequence in epi \( f \) converging to \( \left( {x, f\left( x\right) }\right) \) . Accordingly, by Proposition 5.2, there exists a sequence \( \left( {{v}_{i},{r}_{i}}\right) \) converging to \( \left( {v, r}\right) \) such that \( \left( {{y}_{i}, f\left( {y}_{i}\right) }\right) + {t}_{i}\left( {{v}_{i},{r}_{i}}\right) \in \operatorname{epi}f \) . Thus \[ f\left( {y}_{i}\right) + {t}_{i}{r}_{i} \geq f\left( {{y}_{i} + {t}_{i}{v}_{i}}\right) \] We rewrite this as \[ \frac{f\left( {{y}_{i} + {t}_{i}{v}_{i}}\right) - f\left( {y}_{i}\right) }{{t}_{i}} \leq {r}_{i} \] Taking limits, we obtain \( {f}^{ \circ }\left( {x;v}\right) \leq r \) as desired. We now show that for any \( v \), for any \( \delta \geq 0 \), the point \( \left( {v,{f}^{ \circ }\left( {x;v}\right) + \delta }\right) \) lies in \( {T}_{\text{epi }f}\left( {x, f\left( x\right) }\right) \) ; this will complete the proof of (a). Let \( \left( {{x}_{i},{r}_{i}}\right) \) be any sequence in epi \( f \) converging to \( \left( {x, f\left( x\right) }\right) \), and let \( {t}_{i} \downarrow 0 \) . We must produce a sequence \( \left( {{v}_{i},{s}_{i}}\right) \) converging to \( \left( {v,{f}^{ \circ }\left( {x;v}\right) + \delta }\right) \) with the property that \( \left( {{x}_{i},{r}_{i}}\right) + {t}_{i}\left( {{v}_{i},{s}_{i}}\right) \) lies in epi \( f \) for each \( i \) ; that is, such that \( {r}_{i} + {t}_{i}{s}_{i} \geq \) \( f\left( {{x}_{i} + {t}_{i}{v}_{i}}\right) \) . Let us define \( {v}_{i} = v \) and \[ {s}_{i} \mathrel{\text{:=}} \max \left\{ {{f}^{ \circ }\left( {x;v}\right) + \delta ,\frac{f\left( {{x}_{i} + {t}_{i}v}\right) - f\left( {x}_{i}\right) }{{t}_{i}}}\right\} . \] Observe first that \( {s}_{i} \rightarrow {f}^{ \circ }\left( {x;v}\right) + \delta \), since \[ \mathop{\limsup }\limits_{{i \rightarrow \infty }}\frac{f\left( {{x}_{i} + {t}_{i}v}\right) - f\left( {x}_{i}\right) }{{t}_{i}} \leq {f}^{ \circ }\left( {x;v}\right) . \] We have \[ {r}_{i} + {t}_{i}{s}_{i} \geq {r}_{i} + \left\lbrack {f\left( {{x}_{i} + {t}_{i}v}\right) - f\left( {x}_{i}\right) }\right\rbrack \] and \( {r}_{i} \geq f\left( {x}_{i}\right) \) (since \( \left( {{x}_{i},{r}_{i}}\right) \in \operatorname{epi}f \) ), which together give \[ {r}_{i} + {t}_{i}{s}_{i} \geq f\left( {{x}_{i} + {t}_{i}v}\right) \] showing that \( \left( {{x}_{i} + {t}_{i}v,{r}_{i} + {t}_{i}{s}_{i}}\right) \) belongs to epi \( f \), as required. We turn now to (b). We know that \( \zeta \in \partial f\left( x\right) \) iff \( {f}^{ \circ }\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v \) ; that is, precisely when for any \( v \) and any \( r \geq {f}^{ \circ }\left( {x;v}\right) \) we have \[ \langle \left( {\zeta , - 1}\right) ,\left( {v, r}\right) \rangle \leq 0 \] By (a), this last inequality holds for all the \( \left( {v, r}\right) \) in question iff it holds for all \( \left( {v, r}\right) \in {T}_{\text{epi }f}\left( {x, f\left( x\right) }\right) \) ; that is, precisely when \( \left( {\zeta , - 1}\right) \) lies in the polar of \( {T}_{\text{epi }f}\left( {x, f\left( x\right) }\right) \), namely \( {N}_{\text{epi }f}\left( {x, f\left( x\right) }\right) \) . Remark. One of the principal advantages of the theory of generalized gradients is the complete duality that it induces between tangency and normality, and between functions and sets (via the epigraph, or via the distance function). Note that in developing the theory, we chose the generalized directional derivative as primitive notion, and used it to define the generalized gradient, the tangent cone (via \( {d}_{S} \) ) and then, by polarity, the normal cone. Alternatively, we could choose tangency as a starting point, using the characterization of Proposition 5.2 as a definition, and proceed from there (how?). We do not know in general how to make normality (or generalized gradients, or subdifferentials) the true starting point of the theory, unless the Banach space \( X \) has additional properties, such as in the Hilbert space case considered in Chapter 1. There, normals were the starting point. It is now time to connect the proximal theory of that chapter to the present one. ## 6 Relationship to Proximal Analysis We now suppose that \( X \) is a Hilbert space, so that \( \parallel x\parallel = \langle x, x{\rangle }^{1/2} \) for an inner product \( \langle \cdot , \cdot \rangle \) . In this case, \( \partial f\left( x\right) \) and \( {N}_{S}\left( x\right) \) are generated by weak limits of their proximal counterparts, and are identified with subsets of \( X \) itself: 6.1. Theorem. Let \( X \) be a Hilbert space. (a) If \( f \) is Lipschitz near \( x \), then \[ \partial f\left( x\right) = \overline{\operatorname{co}}\left\{ {\mathop{\operatorname{w-lim}}\limits_{{i \rightarrow \infty }}{\zeta }_{i} : {\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) ,{x}_{i} \rightarrow x}\right\} . \] (b) If \( S \) is a closed subset of \( X \) containing \( x \), then \[ {N}_{S}\left( x\right) = \overline{\operatorname{co}}\left\{ {\underset{i \rightarrow \infty }{\text{ v-lim }}{\zeta }_{i} : {\zeta }_{i} \in {N}_{S}^{P}\left( {x}_{i}\right) ,{x}_{i} \rightarrow x}\right\} . \] Proof. Assertion (a) is an equality between closed convex sets; we will show that the set on the left is contained in the one on the right by verifying the corresponding inequality between support functions. More precisely then, we will show that given any \( v \), there exists a sequence \( {x}_{i} \) converging to \( x \), a sequence \( {\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) \) converging weakly to a limit \( \zeta \), such that \( {f}^{ \circ }\left( {x;v}\right) \leq \langle \zeta, v\rangle \) Let the sequences \( \left\{ {y}_{i}\right\} ,\left\{ {t}_{i}\right\} \) realize \( {f}^{ \circ }\left( {x;v}\right) \) ; i.e., be such that \( {y}_{i} \rightarrow x,{t}_{i} \downarrow 0 \) , and \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{f\left( {{y}_{i} + {t}_{i}v}\right) - f\left( {y}_{i}\right) }{{t}_{i}} = {f}^{ \circ }\left( {x;v}\right) . \] For each \( i \) sufficiently large, \( f \) is Lipschitz on a neighborhood of the line segment \( \left\lbrack {{y}_{i},{y}_{i} + {t}_{i}v}\right\rbrack \), and by the Proximal Mean Value Theorem (Problem 1.11.13) there exists \( {x}_{i} \in \left( {{y}_{i},{y}_{i} + {t}_{i}v}\right) \) and \( {\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) \) such that \[ f\left( {{y}_{i} + {t}_{i}v}\right) - f\left( {y}_{i}\right) < \left\langle {{\zeta }_{i},{t}_{i}v}\right\rangle + \frac{1}{i}. \] Since \( f \) is Lipschitz near \( x \), the sequence \( \left\{ {\zeta }_{i}\right\} \) is bounded, and we may extract a subsequence (without relabeling) that converges weakly to a limit \( \zeta \) . It follows that \( {f}^{ \circ }\left( {x;v}\right) \leq \langle \zeta, v\rangle \), as desired. To prove the opposite inclusion and obtain the result (a), it suffices to prove that any point \( \zeta \) of the form w-lim \( {}_{i \rightarrow \infty }{\zeta }_{i} \), where \( {\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) \) and \( {x}_{i} \rightarrow x \), belongs to \( \partial f\left( x\right) \) (since \( \partial f\left( x\right) \) is itself closed and convex). To see this, observe first that \( {\partial }_{P}f\left( {x}_{i}\right) \subset \partial f\left( {x}_{i}\right) \) (why?); then it suffices to invoke Proposition 1.5(d). Thus (a) is proved. We turn to (b). To show that \( {N}_{S}\left( x\right) \) is contained in the set on the right side of the formula asserted there, it suffices to show that \( \partial {d}_{S}\left( x\right) \) is contained in the right side, in view of Proposition 5.4. In turn, in light of (a), this would follow from proving that any \( \zeta \) of the form w-lim \( {\zeta }_{i} \), where \( {\zeta }_{i} \in \) \( {\partial }_{P}{d}_{S}\left( {x}_{i}\right) \) and \( {x}_{i} \rightarrow x \), belongs to the right side. But this is evident from Theorem 1.6.4. To complete the proof of the theorem, it suffices now to prove that any point \( \zeta \) of the form \( \mathrm{w} \) -lim \( {\zeta }_{i} \), where \( {\zeta }_{i} \in {N}_{S}^{P}\left( {x}_{i}\right) \) and \( {x}_{i} \rightarrow x \), belongs to \( {N}_{S}\left( x\right) \) . We may suppose \( \parallel \zeta \parallel = 1 \) without loss of generality, and (by passing to a subsequence) that \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \rightarrow \lambda > 0 \) . Since \( {\zeta }_{i} \in {N}_{S}^{P}\left( {x}_{i}\right) \), there exists \( {\sigma }_{i} \geq 0 \) such that \[ \left\langle {{\zeta }_{i},{x}^{\prime } - {x}_{i}}\right\rangle \leq {\sigma }_{i}{\begin{Vmatrix}{x}^{\prime } - {x}_{i}\end{Vmatrix}}^{2}\forall {x}^{\prime } \in S. \] Thus the function \( {x}^{\prime } \mapsto \left\langle {-{\zeta }_{i},{x}^{\prime }}\right\rangle + {\sigma }_{i}{\begin{Vmatrix}{x}^{\prime } - {x}_{i}\end{Vmatrix}}^{2} \) attains a minimum over \( S \) at \( {x}^{\prime } = {x}_{i} \) . Note that for \( {x}^{\prime } \) in a neighborhood of \( {x}_{i} \), the function in question has Lipschitz rank \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + 1/i \) . We invoke Proposition 1.6.3 to deduce that locally, \( {x}_{i} \) minimizes the function \[ {x}^{\prime } \mapsto \left\langle {-{\zeta }_{i},{x}^{\prime }}\right\rangle + {\sigma }_{i}{\begin{Vmatrix}{x}^{\prime } - {x}_{i}\end{Vmatrix}}^{2} + \left( {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + \frac{1}{i}}\right) {d}_{S}\left( {x}^{\prime }\right) . \] We now deduce \[ {\zeta }_{i} \in \left( {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + \frac{1}{i}}\right) {\partial }_{P}{d}_{S}\left( {x}_{i}\right) \subset \left( {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + \frac{1}{i}}\right) \partial {d}_{S}\left( {x}_{i}\right) , \] or \( {\zeta }_{i}/\left( {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + 1/i}\right) \in \partial {d}_{S}\left( {x}_{i}\right) \) . But \( {\zeta }_{i}/\left( {\begin{V
1012_(GTM178)Nonsmooth Analysis and Control Theory
32
at \( {x}^{\prime } = {x}_{i} \) . Note that for \( {x}^{\prime } \) in a neighborhood of \( {x}_{i} \), the function in question has Lipschitz rank \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + 1/i \) . We invoke Proposition 1.6.3 to deduce that locally, \( {x}_{i} \) minimizes the function \[ {x}^{\prime } \mapsto \left\langle {-{\zeta }_{i},{x}^{\prime }}\right\rangle + {\sigma }_{i}{\begin{Vmatrix}{x}^{\prime } - {x}_{i}\end{Vmatrix}}^{2} + \left( {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + \frac{1}{i}}\right) {d}_{S}\left( {x}^{\prime }\right) . \] We now deduce \[ {\zeta }_{i} \in \left( {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + \frac{1}{i}}\right) {\partial }_{P}{d}_{S}\left( {x}_{i}\right) \subset \left( {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + \frac{1}{i}}\right) \partial {d}_{S}\left( {x}_{i}\right) , \] or \( {\zeta }_{i}/\left( {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + 1/i}\right) \in \partial {d}_{S}\left( {x}_{i}\right) \) . But \( {\zeta }_{i}/\left( {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} + 1/i}\right) \) converges weakly to \( \zeta /\lambda \), so by Proposition 1.5(d) we obtain \( \zeta \in \lambda \partial {d}_{S}\left( x\right) \subset {N}_{S}\left( x\right) \) (by Proposition 5.4). 6.2. Remark. In terms of the limiting subdifferential \( {\partial }_{L}f \) and the limiting normal cones \( {N}_{S}^{L} \) defined in \( §{1.10} \), the theorem asserts that in a Hilbert space we have \( {N}_{S}\left( x\right) = \overline{\operatorname{co}}{N}_{S}^{L}\left( x\right) \), and, when \( f \) is Lipschitz near \( x,\partial f\left( x\right) = \overline{\operatorname{co}}{\partial }_{L}f\left( x\right) \) . We will establish in Chapter 3 that when \( X = {\mathbb{R}}^{n} \) , a function \( f \) which is regular at \( x \) satisfies \( {\partial }_{L}f\left( x\right) = \partial f\left( x\right) \) . In view of the closed convex hulls that appear in Theorem 6.1, we sometimes adopt the alternative notations \( {\partial }_{C}f,{N}_{S}^{C} \), and \( {T}_{S}^{C} \) for the generalized gradient, normal cone, and tangent cone, particularly when other constructs such as \( {\partial }_{P} \) and \( {\partial }_{L} \) are present simultaneously. We remark as well that \( {\partial }_{C}f \) and \( {T}_{S}^{C} \) are often referred to as Clarke’s generalized gradient and tangent cone, respectively. Theorem 6.1 gives us access in the Hilbert space setting to the results of Chapter 1. It follows, for example (from Exercise 1.10.3(a) and the theorem) that \( {N}_{S}\left( x\right) \neq \{ 0\} \) if \( x \) lies on the boundary of \( S \) and \( X \) is finite dimensional. Another illustration is the following: 6.3. Exercise. Let \( S \in {\mathbb{R}}^{n} \) have the form \[ S \mathrel{\text{:=}} \left\{ {x \in {\mathbb{R}}^{n} : {f}_{i}\left( x\right) = 0, i = 1,2,\ldots, k}\right\} , \] where each \( {f}_{i} : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is \( {C}^{2} \) . Let \( x \in S \), and suppose that the set \( \left\{ {{f}_{i}^{\prime }\left( x\right) : i = 1,2,\ldots, k}\right\} \) is linearly independent. Prove that \[ {N}_{S}^{C}\left( x\right) = \operatorname{span}\left\{ {{f}_{i}^{\prime }\left( x\right) : i = 1,2\ldots, k}\right\} = {N}_{S}^{P}\left( x\right) , \] \[ {T}_{S}^{C}\left( x\right) = \left\{ {v \in {\mathbb{R}}^{n} : \left\langle {{f}_{i}^{\prime }\left( x\right), v}\right\rangle = 0, i = 1,2,\ldots, k}\right\} . \] (Hint. Proposition 1.1.9.) 6.4. Exercise. Confirm the formula of Theorem 6.1(b) in each case of Exercise 5.6. (Note that the formula offers a simpler approach to finding \( {N}_{S}^{C} \) in some cases.) ## 7 The Bouligand Tangent Cone and Regular Sets Let \( S \) be a closed subset of the Banach space \( X, x \in S \) . The Bouligand (or contingent) tangent cone to \( S \) at \( x \), denoted \( {T}_{S}^{B}\left( x\right) \), is defined as follows: \[ {T}_{S}^{B}\left( x\right) \mathrel{\text{:=}} \left\{ {\mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{{x}_{i} - x}{{t}_{i}} : {x}_{i}\overset{S}{ \rightarrow }x,{t}_{i} \downarrow 0}\right\} . \] (Recall that \( {x}_{i}\overset{S}{ \rightarrow }x \) means that \( {x}_{i} \in S\forall i = 1,2,\ldots \), and that \( \mathop{\lim }\limits_{{i \rightarrow \infty }}{x}_{i} = \) \( x \) .) This very natural concept of tangency can be characterized by means of the distance function, as can the tangent cone \( {T}_{S}\left( x\right) \) of \( §5 \) . 7.1. Exercise. (a) Prove that \( v \in {T}_{S}^{B}\left( x\right) \) iff \[ \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{{d}_{S}\left( {x + {tv}}\right) }{t} = 0 \] (b) Deduce that \( {T}_{S}^{C}\left( x\right) \subset {T}_{S}^{B}\left( x\right) \) . (c) Calculate \( {T}_{S}^{B}\left( 0\right) \) for each of the sets \( S \) of Exercise 5.6. Note that in contrast to \( {T}_{S}^{C}\left( x\right) ,{T}_{S}^{B}\left( x\right) \) may not be convex. (d) When \( X \) is a Hilbert space, show that \[ {T}_{S}^{B}\left( x\right) \subset {\left( {N}_{S}^{P}\left( x\right) \right) }^{ \circ }. \] It is natural to inquire into the possible coincidence of \( {T}_{S}^{B}\left( x\right) \) and \( {T}_{S}^{C}\left( x\right) \) . This turns out to be an interesting question meriting a definition. We will say that a set \( S \) is regular at \( x \in S \) provided that \( {T}_{S}^{B}\left( x\right) = {T}_{S}^{C}\left( x\right) \) . 7.2. Exercise. Show that a convex set \( S \) is regular at each of its points. (Hint. \( {d}_{S}\left( \cdot \right) \) is convex.) We have used the term "regular" both for functions (§4) and now for sets; here is why the use of the same term is appropriate: 7.3. Proposition. Let \( f \) be Lipschitz near \( x \) . Then \( f \) is regular at \( x \) iff epi \( f \) is regular at \( \left( {x, f\left( x\right) }\right) \) . Proof. Let us suppose first that \( f \) is regular at \( x \) . We wish to show that \( {T}_{\text{epi }}^{B}f\left( {x, f\left( x\right) }\right) \) and \( {T}_{\text{epi }}^{C}f\left( {x, f\left( x\right) }\right) \) agree; equivalently, that an element \( \left( {v, r}\right) \) of \( {T}_{\text{epi }f}^{B}\left( {x, f\left( x\right) }\right) \) must lie in \( {T}_{\text{epi }f}^{C}\left( {x, f\left( x\right) }\right) = \operatorname{epi}{f}^{ \circ }\left( {x; \cdot }\right) \) (by Theorem 5.7), which is to say that we have \( {f}^{ \circ }\left( {x;v}\right) \leq r \) . Since \( {f}^{\prime }\left( {x;v}\right) = {f}^{ \circ }\left( {x;v}\right) \), it suffices to produce a sequence \( {t}_{i} \downarrow 0 \) such that \[ \mathop{\liminf }\limits_{{i \rightarrow \infty }}\frac{f\left( {x + {t}_{i}v}\right) - f\left( x\right) }{{t}_{i}} \leq r. \] Now by Exercise 7.1(a), we have \[ \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{{d}_{\text{epi }f}\left( {\left( {x, f\left( x\right) }\right) + t\left( {v, r}\right) }\right) }{t} \leq 0. \] That this condition implies the existence of the stated sequence \( {t}_{i} \) is easy to establish. For the converse, suppose now that epi \( f \) is regular at \( \left( {x, f\left( x\right) }\right) \) . To establish that \( f \) is regular at \( x \), we will show that for any \( v \in X \), we have \[ \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{f\left( {x + {tv}}\right) - f\left( x\right) }{t} \geq {f}^{ \circ }\left( {x;v}\right) \] (why does this imply that \( f \) is regular at \( x \) ?). In turn, this follows if we show that for any sequence \( {t}_{i} \downarrow 0 \), we have \[ \lambda \mathrel{\text{:=}} \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{f\left( {x + {t}_{i}v}\right) - f\left( x\right) }{{t}_{i}} \geq {f}^{ \circ }\left( {x;v}\right) \] whenever this limit \( \lambda \) exists. But we have \[ \left( {v,\lambda }\right) = \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{\left( {x + {t}_{i}v, f\left( {x + {t}_{i}v}\right) }\right) - \left( {x, f\left( x\right) }\right) }{{t}_{i}}, \] which shows that \( \left( {v,\lambda }\right) \in {T}_{\text{epi }f}^{B}\left( {x, f\left( x\right) }\right) = {T}_{\text{epi }f}^{C}\left( {x, f\left( x\right) }\right) \) (since epi \( f \) is regular at \( \left( {x, f\left( x\right) }\right) ) = \operatorname{epi}{f}^{ \circ }\left( {x; \cdot }\right) \) (by Theorem 5.7). It follows that \( {f}^{ \circ }\left( {x;v}\right) \leq \lambda \), as required. The principal role of set regularity is to provide "more exact" estimates in formulas for tangent and normal cones, just as functional regularity led to sharper results for generalized gradients. We illustrate this in the following result, which will be extended to the case of mixed equalities and inequalities in Chapter 3. 7.4. Proposition. Let \( S \) be a subset of \( {\mathbb{R}}^{n} \) given as follows: \[ S \mathrel{\text{:=}} \left\{ {x \in {\mathbb{R}}^{n} : {f}_{j}\left( x\right) = 0, j = 1,2,\ldots, k}\right\} , \] where each \( {f}_{j} : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is locally Lipschitz and admits one-sided directional derivatives \( {f}_{j}^{\prime }\left( {x;v}\right) \) for each \( v \) . Then \[ {T}_{S}^{B}\left( x\right) \subset \left\{ {v \in {\mathbb{R}}^{n} : {f}_{j}^{\prime }\left( {x;v}\right) = 0, j = 1,2,\ldots, k}\right\} . \] If in addition each \( {f}_{j} \) is \( {C}^{1} \) near \( x \) and the vectors \( {\left\{ {f}_{j}^{\prime }\left( x\right) \right\} }_{j = 1}^{k} \) are linearly independent, then \[ \left\{ {v : \left\langle {{f}_{j}^{\prime }\left( x\right), v}\right\rangle = 0, j = 1,2,\ldots, k}\right\} \subset {T}_{S}^{C}\left( x\right) ; \] in this latter case, equality holds in both estimates, and the set \( S \) is regular at \( x \) . Proof. Any \( v \in {T}_{S}^{B}\left( x\right) \) is of the form \( \mathop{\lim }\limits_{{i \rightarrow \infty }}\left( {{x}_{i} - x}\right) /\left( {t}_{i}\right) \), where \( {x}_{i}\overset{S}{ \rightarrow }x \) and \( {t}_{i} \downarrow 0 \) . Then \( {f}_{j}\left( {x}_{i}\right) = 0\left( {j = 1,2,\ldots, k}\right) \), and we deduce \[
1012_(GTM178)Nonsmooth Analysis and Control Theory
33
ves \( {f}_{j}^{\prime }\left( {x;v}\right) \) for each \( v \) . Then \[ {T}_{S}^{B}\left( x\right) \subset \left\{ {v \in {\mathbb{R}}^{n} : {f}_{j}^{\prime }\left( {x;v}\right) = 0, j = 1,2,\ldots, k}\right\} . \] If in addition each \( {f}_{j} \) is \( {C}^{1} \) near \( x \) and the vectors \( {\left\{ {f}_{j}^{\prime }\left( x\right) \right\} }_{j = 1}^{k} \) are linearly independent, then \[ \left\{ {v : \left\langle {{f}_{j}^{\prime }\left( x\right), v}\right\rangle = 0, j = 1,2,\ldots, k}\right\} \subset {T}_{S}^{C}\left( x\right) ; \] in this latter case, equality holds in both estimates, and the set \( S \) is regular at \( x \) . Proof. Any \( v \in {T}_{S}^{B}\left( x\right) \) is of the form \( \mathop{\lim }\limits_{{i \rightarrow \infty }}\left( {{x}_{i} - x}\right) /\left( {t}_{i}\right) \), where \( {x}_{i}\overset{S}{ \rightarrow }x \) and \( {t}_{i} \downarrow 0 \) . Then \( {f}_{j}\left( {x}_{i}\right) = 0\left( {j = 1,2,\ldots, k}\right) \), and we deduce \[ {f}_{j}^{\prime }\left( {x;v}\right) = \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{{f}_{j}\left( {x + {t}_{i}v}\right) - {f}_{j}\left( x\right) }{{t}_{i}} = \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{{f}_{j}\left( {x + {t}_{i}v}\right) }{{t}_{i}} \] \[ = \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{{f}_{j}\left( {x}_{i}\right) }{{t}_{i}} = 0 \] since \( \left| {{f}_{j}\left( {x + {t}_{i}v}\right) - {f}_{j}\left( {x}_{i}\right) }\right| /{t}_{i} \leq K\begin{Vmatrix}{x + {t}_{i}v - {x}_{i}}\end{Vmatrix}/{t}_{i} \rightarrow 0 \), where \( K \) is a Lipschitz constant for \( {f}_{j} \) . This gives the stated estimate for \( {T}_{S}^{B}\left( x\right) \) . Now let us posit the additional hypotheses on the \( {f}_{j} \) . Then Proposition 1.9(a) applies, for all \( {x}^{\prime } \) sufficiently near \( x \), and yields \[ {N}_{S}^{P}\left( {x}^{\prime }\right) \subset \operatorname{span}{\left\{ {f}_{j}^{\prime }\left( {x}^{\prime }\right) \right\} }_{j = 1}^{k}. \] We can now appeal to Theorem 6.1(b) to obtain \[ {N}_{S}^{C}\left( x\right) \subset \operatorname{span}{\left\{ {f}_{j}^{\prime }\left( x\right) \right\} }_{j = 1}^{k}. \] Polarity transforms this estimate for \( {N}_{S}^{C}\left( x\right) \) to precisely the one for \( {T}_{S}^{C}\left( x\right) \) that is claimed. Since \( {T}_{S}^{C}\left( x\right) \subset {T}_{S}^{B}\left( x\right) \) is always true, equality holds in both estimates of the proposition, and so \( S \) is regular at \( x \) . ## 8 The Gradient Formula in Finite Dimensions The celebrated theorem of Rademacher asserts that if a function \( f : {\mathbb{R}}^{n} \rightarrow \) \( \mathbb{R} \) is Lipschitz on an open set \( U \), then it is differentiable almost everywhere (a.e.) on \( U \) (in the sense of Lebesgue measure); we prove this theorem in Chapter 3. It turns out that the derivative of \( f \) can be used to generate its generalized gradient, as depicted in the following formula, one of the most useful computational tools in nonsmooth analysis. It shows that in \( {\mathbb{R}}^{n},\partial f\left( x\right) \) can be generated by the values of \( \nabla f\left( {x}^{\prime }\right) \) at nearby points \( {x}^{\prime } \) at which \( {f}^{\prime }\left( {x}^{\prime }\right) \) exists, and furthermore that points \( {x}^{\prime } \) belonging to any prescribed set of measure zero can be ignored in the construction without changing the result. This latter aspect of \( \partial f\left( x\right) \) is referred to as being "blind to sets of measure zero." 8.1. Theorem (Generalized Gradient Formula). Let \( x \in {\mathbb{R}}^{n} \), and let \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be Lipschitz near \( x \) . Let \( \Omega \) be any subset of zero measure in \( {\mathbb{R}}^{n} \) , and let \( {\Omega }_{f} \) be the set of points in \( {\mathbb{R}}^{n} \) at which \( f \) fails to be differentiable. Then \[ \partial f\left( x\right) \mathrel{\text{:=}} \operatorname{co}\left\{ {\lim \nabla f\left( {x}_{i}\right) : {x}_{i} \rightarrow x,{x}_{i} \notin \Omega ,{x}_{i} \notin {\Omega }_{f}}\right\} . \] Proof. The meaning of the formula is the following: consider any sequence \( \left\{ {x}_{i}\right\} \) converging to \( x \) while avoiding both \( \Omega \) and points at which \( f \) is not differentiable, and such that the sequence \( \left\{ {\nabla f\left( {x}_{i}\right) }\right\} \) converges; then the convex hull of all such limits is \( \partial f\left( x\right) \) . As usual in a Hilbert space setting, we identify \( \partial f\left( x\right) \) with a subset of \( {\mathbb{R}}^{n} \), the space itself. Let us note to begin with that there are "plenty" of sequences \( {x}_{i} \) which converge to \( x \) and avoid \( \Omega \cup {\Omega }_{f} \), since the latter has measure 0 near \( x \) . Further, because \( \partial f \) is locally bounded near \( x \) (Proposition 1.5) and \( \nabla f\left( {x}_{i}\right) \) belongs to \( \partial f\left( {x}_{i}\right) \) for each \( i \) (Proposition 3.1) the sequence \( \left\{ {\nabla f\left( {x}_{i}\right) }\right\} \) admits a convergent subsequence by the Bolzano-Weierstrass Theorem. The limit of any such sequence must belong to \( \partial f\left( x\right) \) by the closure property of \( \partial f \) proved in Proposition 1.5. It follows that the set \[ \left\{ {\lim \nabla f\left( {x}_{i}\right) : {x}_{i} \rightarrow x,{x}_{i} \notin \Omega \cup {\Omega }_{f}}\right\} \] is contained in \( \partial f\left( x\right) \) and is nonempty and bounded, and in fact compact, since it is rather obviously closed. Since \( \partial f\left( x\right) \) is convex, we deduce that the left-hand side of the formula asserted by the theorem contains the right. Now, the convex hull of a compact set in \( {R}^{n} \) is compact, so to complete the proof we need only show that the support function of the left-hand side (i.e., \( {f}^{ \circ }\left( {x; \cdot }\right) \) ) never exceeds that of the right. This is what the following lemma does: Lemma. For any \( v \neq 0 \) in \( {R}^{n} \), for any \( \varepsilon > 0 \), we have \[ {f}^{ \circ }\left( {x;v}\right) - \varepsilon \leq \lim \sup \left\{ {\nabla f\left( y\right) \cdot v : y \rightarrow x, y \notin \Omega \cup {\Omega }_{f}}\right\} . \] To prove this, let the right-hand side be \( \alpha \) . Then by definition, there is a \( \delta > 0 \) such that the conditions \[ y \in x + {\delta B},\;y \notin \Omega \cup {\Omega }_{f}, \] imply \( \nabla f\left( y\right) \cdot v \leq \alpha + \varepsilon \) . We also choose \( \delta \) small enough so that \( f \) is Lipschitz on \( B\left( {x;\delta }\right) \) and \( \Omega \cup {\Omega }_{f} \) has measure 0 in \( x + {\delta B} \) . Now consider the line segments \( {L}_{y} = \{ y + {tv} : 0 < t < \delta /\left( {2\left| v\right| }\right) \} \) . Since \( \Omega \cup {\Omega }_{f} \) has measure 0 in \( x + {\delta B} \), it follows from Fubini’s Theorem that for almost every \( y \) in \( x + \left( {\delta /2}\right) B \), the line segment \( {L}_{y} \) meets \( \Omega \cup {\Omega }_{f} \) in a set of 0 one-dimensional measure. Let \( y \) be any point in \( x + \left( {\delta /2}\right) B \) having this property, and let \( t \) lie in \( \left( {0,\delta /\left( {2\left| v\right| }\right) }\right) \) . Then \[ f\left( {y + {tv}}\right) - f\left( y\right) = {\int }_{0}^{t}\nabla f\left( {y + {sv}}\right) \cdot {vds} \] since \( {f}^{\prime } \) exists a.e. on \( {L}_{y} \) . Since we have \( \parallel y + {sv} - x\parallel < \delta \) for \( 0 < s < t \), it follows that \( \nabla f\left( {y + {sv}}\right) \cdot v \leq \alpha + \varepsilon \), whence \[ f\left( {y + {tv}}\right) - f\left( y\right) \leq t\left( {\alpha + \varepsilon }\right) . \] Since this is true for all \( y \) within \( \delta /2 \) of \( x \) except those in a set of measure 0, and for all \( t \) in \( \left( {0,\delta /\left( {2\left| v\right| }\right) }\right) \), and since \( f \) is continuous, it is in fact true for all such \( y \) and \( t \) . We deduce \[ {f}^{ \circ }\left( {x;v}\right) \leq \alpha + \varepsilon \] which completes the proof. 8.2. Corollary. \[ {f}^{ \circ }\left( {x;v}\right) = \mathop{\limsup }\limits_{{y \rightarrow x}}\left\{ {\nabla f\left( y\right) \cdot v : y \notin \Omega \cup {\Omega }_{f}}\right\} . \] 8.3. Exercise. Use the theorem to calculate \( \partial f\left( {0,0}\right) \) where \( f \) on \( {R}^{2} \) is given by \[ f\left( {x, y}\right) = \max \{ \min \left\lbrack {x, - y}\right\rbrack, y - x\} \] as follows. Define \[ {C}_{1} = \{ \left( {x, y}\right) : y \leq {2x}\text{ and }y \leq - x\} \] \[ {C}_{2} = \{ \left( {x, y}\right) : y \leq x/2\text{ and }y \geq - x\} \] \[ {C}_{3} = \{ \left( {x, y}\right) : y \geq {2x}\text{ or }y \geq x/2\} . \] Then \( {C}_{1} \cup {C}_{2} \cup {C}_{3} = {R}^{2} \), and we have \[ f\left( {x, y}\right) = \begin{cases} x & \text{ for }\left( {x, y}\right) \in {C}_{1}, \\ - y & \text{ for }\left( {x, y}\right) \in {C}_{2}, \\ y - x & \text{ for }\left( {x, y}\right) \in {C}_{3}. \end{cases} \] Note that the boundaries of these three sets form a set \( \Omega \) of measure 0, and that if \( \left( {x, y}\right) \) does not lie in \( \Omega \), then \( f \) is differentiable and \( \nabla f\left( {x, y}\right) \) is one of the points \( \left( {1,0}\right) ,\left( {0, - 1}\right) \), or \( \left( {-1,1}\right) \) . Then \( \partial f\left( {0,0}\right) \) is the triangle obtained as the convex hull of these three points. The derivative of the distance function \( {d}_{S}\left( \cdot \right) \) of a closed set \( S \) in \( {\mathbb{R}}^{n} \) admits an interpretation along the lines of its proximal subgradients (Theorem 1.6.1). 8.4. Proposition. Let \( {d}_{S}^{\prime }\left( x\right) \) exist and be different from 0 . Then \( x \notin S \) , \( {\operatorname{proj}}_{S}\left( x\right) \) is a singleton \( \{ s\} \), and \( \nabla {
1012_(GTM178)Nonsmooth Analysis and Control Theory
34
ht) \in {C}_{1}, \\ - y & \text{ for }\left( {x, y}\right) \in {C}_{2}, \\ y - x & \text{ for }\left( {x, y}\right) \in {C}_{3}. \end{cases} \] Note that the boundaries of these three sets form a set \( \Omega \) of measure 0, and that if \( \left( {x, y}\right) \) does not lie in \( \Omega \), then \( f \) is differentiable and \( \nabla f\left( {x, y}\right) \) is one of the points \( \left( {1,0}\right) ,\left( {0, - 1}\right) \), or \( \left( {-1,1}\right) \) . Then \( \partial f\left( {0,0}\right) \) is the triangle obtained as the convex hull of these three points. The derivative of the distance function \( {d}_{S}\left( \cdot \right) \) of a closed set \( S \) in \( {\mathbb{R}}^{n} \) admits an interpretation along the lines of its proximal subgradients (Theorem 1.6.1). 8.4. Proposition. Let \( {d}_{S}^{\prime }\left( x\right) \) exist and be different from 0 . Then \( x \notin S \) , \( {\operatorname{proj}}_{S}\left( x\right) \) is a singleton \( \{ s\} \), and \( \nabla {d}_{S}\left( x\right) = \left( {x - s}\right) /\parallel x - s\parallel \) . Proof. If \( x \) lies in \( S \), then \( {d}_{S}^{\prime }\left( x\right) = 0 \) since \( {d}_{S} \) has a minimum at \( x \) ; thus \( x \notin \) \( S \) . Let \( s \in {\operatorname{proj}}_{S}\left( x\right) \) . We will show that \( \nabla {d}_{S}\left( x\right) = \left( {x - s}\right) /\parallel x - s\parallel \), thereby establishing also that \( {\operatorname{proj}}_{S}\left( x\right) \) contains no other points. For \( t \in \left( {0,1}\right) \), the closest point in \( S \) to \( x + t\left( {s - x}\right) \) is \( s \) (see Proposition 1.1.3), whence \[ {d}_{S}\left( {x + t\left( {s - x}\right) }\right) = \left( {1 - t}\right) \parallel x - s\parallel . \] Subtracting \( {d}_{S}\left( x\right) = \parallel x - s\parallel \) from both sides, dividing by \( t \), and letting \( t \downarrow 0 \) produces \[ {d}_{S}^{\prime }\left( {x;s - x}\right) = \left\langle {\nabla {d}_{S}\left( x\right), s - x}\right\rangle = - \parallel s - x\parallel . \] Now \( \begin{Vmatrix}{\nabla {d}_{S}\left( x\right) }\end{Vmatrix} \leq 1 \), since \( {d}_{S} \) is Lipschitz of rank 1, so the last equation implies \( \nabla {d}_{S}\left( x\right) = \left( {x - s}\right) /\parallel x - s\parallel \) . 2. Generalized Gradients in Banach Space 8.5. Exercise. (a) Let \( x \in \operatorname{bdry}\left( S\right) \) . Use Theorem 8.1, Exercise 5.4, and Proposition 8.4 to deduce the following formula for \( {N}_{S}\left( x\right) \) when \( S \) is a closed subset of \( {\mathbb{R}}^{n} \) and \( \Omega \) is any set of measure 0 : \[ {N}_{S}\left( x\right) = \overline{\operatorname{co}}\left\{ {\mathop{\lim }\limits_{{i \rightarrow \infty }}{\lambda }_{i}\left( {{x}_{i} - {s}_{i}}\right) : {\lambda }_{i} \geq 0,{x}_{i} \notin \Omega \cup S,}\right. \] \[ \left. {{x}_{i} \rightarrow x,{\operatorname{proj}}_{S}\left( {x}_{i}\right) = \left\{ {s}_{i}\right\} }\right\} . \] It follows that \( {N}_{S}\left( x\right) \) contains nonzero elements. (b) Use this characterization to calculate the normal cones of Exercise 5.6. ## 9 Problems on Chapter 2 9.1. A function \( f : X \rightarrow \mathbb{R} \) is said to be strictly (Hadamard) differentiable at \( x \) if there exists an element \( \zeta \in {X}^{ * } \) such that for each \( v \) in \( X \) we have \[ \mathop{\lim }\limits_{\substack{{{x}^{\prime } \rightarrow x} \\ {t \downarrow 0} }}\frac{f\left( {{x}^{\prime } + {tv}}\right) - f\left( {x}^{\prime }\right) }{t} = \langle \zeta, v\rangle \] and provided the convergence is uniform for \( v \) in compact sets. (a) \( f \) is strictly differentiable at \( x \) iff \( f \) is Lipschitz near \( x \) and \( \partial f\left( x\right) \) is a singleton. (b) If \( f \) is regular at \( x \) and \( {f}^{\prime }\left( x\right) \) exists, then \( f \) is strictly differentiable at \( x \) . (c) If \( f \) is regular at \( x \) and \( {f}^{\prime }\left( x\right) \) exists, and if \( g \) is Lipschitz near \( x \), then \[ \partial \left( {f + g}\right) \left( x\right) = \left\{ {{f}^{\prime }\left( x\right) }\right\} + \partial g\left( x\right) . \] (d) If \( X \) is finite dimensional and \( U \) is an open subset of \( X \), then \( f \in \) \( {C}^{1}\left( U\right) \) iff \( f \) is locally Lipschitz on \( U \) and \( \partial f\left( x\right) \) is a singleton for each \( x \in U \) . 9.2. Let \( \varphi : \left\lbrack {0,1}\right\rbrack \rightarrow \mathbb{R} \) belong to \( {L}^{\infty }\left\lbrack {0,1}\right\rbrack \) and define a (Lipschitz) function \( f : \left\lbrack {0,1}\right\rbrack \rightarrow \mathbb{R} \) via \( f\left( x\right) \mathrel{\text{:=}} {\int }_{0}^{x}\varphi \left( t\right) {dt} \) . Prove that \( \partial f\left( x\right) \) is the interval \( \left\lbrack {{\varphi }^{ - }\left( x\right) ,{\varphi }^{ + }\left( x\right) }\right\rbrack \), where \( {\varphi }^{ - }\left( x\right) \) and \( {\varphi }^{ + }\left( x\right) \) are the essential infimum and essential supremum, respectively, of \( \varphi \) at \( x \) ; thus, for instance, \[ {\varphi }^{ + }\left( x\right) \mathrel{\text{:=}} \inf \left\{ {M : \exists \varepsilon > 0 \ni \varphi \left( {x}^{\prime }\right) \leq M\text{ a.e. for }{x}^{\prime } \in \left\lbrack {x - \varepsilon, x + \varepsilon }\right\rbrack }\right\} . \] 9.3. For \( f \) as in Problem 9.2, show that there is a point \( z \in \left( {0,1}\right) \) such that for any \( \delta > 0 \), there exist \( x \) and \( y \) within \( \delta \) of \( z \) such that \[ \varphi \left( x\right) - \delta \leq {\int }_{0}^{1}\varphi \left( t\right) {dt} \leq \varphi \left( y\right) + \delta \] (When \( \varphi \) is continuous this holds for \( \delta = 0 \) : the Intermediate Value Theorem.) 9.4. Let \( X = {X}_{1} \times {X}_{2} \), where \( {X}_{1} \) and \( {X}_{2} \) are Banach spaces, and let \( f\left( {{x}_{1},{x}_{2}}\right) \) on \( X \) be Lipschitz near \( \left( {{x}_{1},{x}_{2}}\right) \) . We denote by \( {\partial }_{1}f\left( {{x}_{1},{x}_{2}}\right) \) the (partial) generalized gradient of \( f\left( {\cdot ,{x}_{2}}\right) \) at \( {x}_{1} \), and similarly for \( {\partial }_{2}f \) . (a) Let \( {\pi }_{1}\partial f\left( {{x}_{1},{x}_{2}}\right) \) be the set \[ \left\{ {{\zeta }_{1} \in {X}_{1}^{ * } : \text{ for some }{\zeta }_{2} \in {X}_{2}^{ * },\left( {{\zeta }_{1},{\zeta }_{2}}\right) \in \partial f\left( {{x}_{1},{x}_{2}}\right) }\right\} . \] Prove that \( {\partial }_{1}f\left( {{x}_{1},{x}_{2}}\right) \subset {\pi }_{1}\partial f\left( {{x}_{1},{x}_{2}}\right) \), with equality when \( f \) is regular at \( \left( {{x}_{1},{x}_{2}}\right) \) . (Hint. Consider Theorem 3.2 and \( F\left( x\right) = \left( {x,{x}_{2}}\right) \) .) (b) Calculate \( {\partial }_{1}f\left( {0,0}\right) ,{\partial }_{2}f\left( {0,0}\right) \) for the function \( f \) of Exercise 8.3. Note that \[ {\partial }_{1}f\left( {0,0}\right) \times {\partial }_{2}f\left( {0,0}\right) ⊄ \partial f\left( {0,0}\right) ⊄ {\partial }_{1}f\left( {0,0}\right) \times {\partial }_{2}f\left( {0,0}\right) . \] 9.5. Let \( X \) be a Banach space, and let \( f : X \rightarrow \mathbb{R} \) be Lipschitz near \( x \) . Suppose that \( x \) minimizes \( f \) over a closed set \( S \) . Then \( 0 \in \partial f\left( x\right) + {N}_{S}\left( x\right) \) . (Hint. Use exact penalization as in Proposition 1.6.3.) 9.6. Let \( f : X \rightarrow \mathbb{R} \) be Lipschitz near \( x \), where \( X \) is a Hilbert space, and suppose \( 0 \notin \partial f\left( x\right) \) . Let \( \zeta \) be the element of least norm in \( \partial f\left( x\right) \) . Prove that \( - \zeta \) is a "direction of descent" in the following sense: for all \( t > 0 \) sufficiently small we have \[ f\left( {x - {t\zeta }}\right) \leq f\left( x\right) - t\frac{\parallel \zeta {\parallel }^{2}}{2}. \] 9.7. Show by an example in one dimension that unlike \( \partial f,{\partial }_{L}f \) is not "blind to sets of measure 0"; that is, we cannot neglect an arbitrary set of measure 0 in the formula \[ {\partial }_{L}f\left( x\right) = \left\{ {\mathop{\lim }\limits_{{i \rightarrow \infty }}{\zeta }_{i} : {\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) ,{x}_{i}\overset{f}{ \rightarrow }x}\right\} \] even when \( f \) is Lipschitz. 9.8. For \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) locally Lipschitz, let \[ {Df}\left( {x;v}\right) \mathrel{\text{:=}} \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{f\left( {x + {tv}}\right) - f\left( x\right) }{t} \] denote the Dini subderivate. Prove that \[ {f}^{ \circ }\left( {x;v}\right) = \mathop{\limsup }\limits_{{y \rightarrow x}}{Df}\left( {y;v}\right) \] 9.9. (Analysis of a Line-Fitting Problem.) We wish to minimize over \( {\mathbb{R}}^{2} \) the function \[ f\left( {\alpha ,\beta }\right) \mathrel{\text{:=}} \left| {{\alpha N} + \beta }\right| + \mathop{\sum }\limits_{{i = 0}}^{{N - 1}}\left| {{\alpha i} + \beta - i}\right| \] where \( N \) is a given positive integer. (We interpret this as finding the best line to fit the given data points \( \left( {1,1}\right) ,\left( {2,2}\right) ,\ldots ,\left( {N - 1, N - 1}\right) \), and \( \left( {N,0}\right) \) .) (a) Let \( g\left( {\alpha ,\beta }\right) \mathrel{\text{:=}} \left| {{\alpha c} + \beta - k}\right| \) for given \( c \) and \( k \) . Prove that \[ \partial g\left( {\alpha ,\beta }\right) = \left\{ \begin{array}{ll} \{ \left( {c,1}\right) \} & \text{ if }{\alpha c} + \beta - k > 0 \\ \{ \left( {-c, - 1}\right) \} & \text{ if }{\alpha c} + \beta - k < 0 \\ \{ \lambda \left( {c,1}\right) : \left| \lambda \right| \leq 1\} & \text{ if }{\alpha c} + \beta - k = 0 \end{array}\right. \] (b) Use (a) to interpret the necessary condition \( 0 \in \partial f\left( {\alpha ,\beta }\right) \) in the form \[ 0 = {\lambda }_{N}\left( {N,1}\right) + \mathop{\sum }\limits_{{i = 0}}^{{N - 1}}{\lambda }_{i}\left( {i,1}\right) \] for
1012_(GTM178)Nonsmooth Analysis and Control Theory
35
lpha i} + \beta - i}\right| \] where \( N \) is a given positive integer. (We interpret this as finding the best line to fit the given data points \( \left( {1,1}\right) ,\left( {2,2}\right) ,\ldots ,\left( {N - 1, N - 1}\right) \), and \( \left( {N,0}\right) \) .) (a) Let \( g\left( {\alpha ,\beta }\right) \mathrel{\text{:=}} \left| {{\alpha c} + \beta - k}\right| \) for given \( c \) and \( k \) . Prove that \[ \partial g\left( {\alpha ,\beta }\right) = \left\{ \begin{array}{ll} \{ \left( {c,1}\right) \} & \text{ if }{\alpha c} + \beta - k > 0 \\ \{ \left( {-c, - 1}\right) \} & \text{ if }{\alpha c} + \beta - k < 0 \\ \{ \lambda \left( {c,1}\right) : \left| \lambda \right| \leq 1\} & \text{ if }{\alpha c} + \beta - k = 0 \end{array}\right. \] (b) Use (a) to interpret the necessary condition \( 0 \in \partial f\left( {\alpha ,\beta }\right) \) in the form \[ 0 = {\lambda }_{N}\left( {N,1}\right) + \mathop{\sum }\limits_{{i = 0}}^{{N - 1}}{\lambda }_{i}\left( {i,1}\right) \] for certain constants \( {\lambda }_{i}\left( {i = 0,1,\ldots, N}\right) \) with values depending on \( \left( {\alpha ,\beta }\right) \) . (c) For \( N \geq 3 \), invoke the convexity of \( f \) to deduce that \( \alpha = 1,\beta = 0 \) minimizes \( f \) . (d) If \( N > 4 \), no other \( \left( {\alpha ,\beta }\right) \) satisfy the necessary condition of (b). (Thus the anomalous point \( \left( {N,0}\right) \) is ignored by the nonsmooth criterion, which is not the case in the usual least squares fit.) 9.10. Let \( S \) in \( {\mathbb{R}}^{n} \) be of the form \( \{ x : f\left( x\right) \leq 0\} \), where \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is locally Lipschitz. Suppose that \( x \in S \), and \( 0 \notin \partial f\left( x\right) \) . Prove that \( \operatorname{int}{T}_{S}\left( x\right) \neq \varnothing \) , and that \[ {T}_{S}\left( x\right) \supset \left\{ {v \in {\mathbb{R}}^{n} : {f}^{ \circ }\left( {x;v}\right) \leq 0}\right\} . \] Show that equality holds if \( f \) is regular at \( x \), in which case \( S \) is regular at \( x \) . (Remark. Sets which admit tangent cones having nonempty interior have useful properties; they will be studied in Chapter 3.) 9.11. We will prove a formula for the tangent and normal cones of the intersection of two closed subsets \( {S}_{1} \) and \( {S}_{2} \) of \( {\mathbb{R}}^{n} \) . (We remind the reader that the limiting normal cone \( {N}_{S}^{L} \) was introduced in \( §{1.10} \), and the alternative notation \( {N}_{S}^{C} \) and \( {T}_{S}^{C} \) for \( {N}_{S} \) and \( {T}_{S} \) in Remark 6.2.) (a) Prove that \( {T}_{C}^{B}\left( x\right) \subset {T}_{S}^{B}\left( x\right) \) if \( C \subset S \) . (Show that this monotonicity property fails for \( {T}^{C} \) .) Deduce that \[ {T}_{{S}_{1} \cap {S}_{2}}^{B}\left( x\right) \subset {T}_{{S}_{1}}^{B}\left( x\right) \cap {T}_{{S}_{2}}^{B}\left( x\right) . \] (b) Note that \( {T}_{S}^{B}{\left( x\right) }^{ \circ } \subset {N}_{S}^{C}\left( x\right) \) generally, and deduce \[ {N}_{{S}_{1} \cap {S}_{2}}^{C}\left( x\right) \supset {T}_{{S}_{1}}^{B}{\left( x\right) }^{ \circ } + {T}_{{S}_{2}}^{B}{\left( x\right) }^{ \circ }. \] (c) Suppose henceforth that \( {N}_{{S}_{1}}^{L}\left( x\right) \cap \left( {-{N}_{{S}_{2}}^{L}\left( x\right) }\right) = \{ 0\} \) (a condition called transversality). Prove that \( {N}_{{S}_{1} \cap {S}_{2}}^{L}\left( x\right) \subset {N}_{{S}_{1}}^{L}\left( x\right) + {N}_{{S}_{2}}^{L}\left( x\right) \) . (Hint. See the argument that proved Proposition 1.10.1.) Give a counterexample to this conclusion when transversality fails. (d) Theorem. If transversality holds at \( x \), then \[ {N}_{{S}_{1} \cap {S}_{2}}^{C}\left( x\right) \subset {N}_{{S}_{1}}^{C}\left( x\right) + {N}_{{S}_{2}}^{C}\left( x\right) ,\;{T}_{{S}_{1} \cap {S}_{2}}^{C}\left( x\right) \supset {T}_{{S}_{1}}^{C}\left( x\right) \cap {T}_{{S}_{2}}^{C}\left( x\right) , \] and equality holds if both \( {S}_{1} \) and \( {S}_{2} \) are regular at \( x \), in which case \( {S}_{1} \cap {S}_{2} \) is also regular at \( x \) . 9.12. Let \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be Lipschitz near \( x \) . (a) \( {f}^{\prime }\left( x\right) \) exists iff \( {T}_{\operatorname{gr}f}^{B}\left( {x, f\left( x\right) }\right) \) is a hyperplane. (b) \( f \) is strictly differentiable at \( x \) iff \( {T}_{\operatorname{gr}f}^{C}\left( {x, f\left( x\right) }\right) \) is a hyperplane. (c) \( f \) is strictly differentiable at \( x \) iff \( {T}_{\text{epi }f}^{C}\left( {x, f\left( x\right) }\right) \) is a half-space. (Hint. Theorem 5.7.) (d) The fact that \( {T}_{\text{epi }f}^{B}\left( {x, f\left( x\right) }\right) \) is a half-space does not imply that \( {f}^{\prime }\left( x\right) \) exists. 9.13. (Danskin’s Theorem.) Let a continuous function \( g : {\mathbb{R}}^{n} \times M \rightarrow \mathbb{R} \) be given, where \( M \) is a compact metric space. We suppose that for a neighborhood \( \Omega \) of a given point \( x \in {\mathbb{R}}^{n} \), the derivative \( {g}_{x}\left( {{x}^{\prime }, u}\right) \) exists and is continuous (jointly) for \( \left( {{x}^{\prime }, u}\right) \in \Omega \times M \) . We set \[ f\left( {x}^{\prime }\right) \mathrel{\text{:=}} \mathop{\max }\limits_{{u \in M}}g\left( {{x}^{\prime }, u}\right) \] Then \( f \) is Lipschitz and regular near \( x \), and we have \[ {\partial }_{C}f\left( x\right) = \operatorname{co}\left\{ {{g}_{x}\left( {x, u}\right) : u \in M\left( x\right) }\right\} \] where \( M\left( x\right) \mathrel{\text{:=}} \{ u \in M : g\left( {x, u}\right) = f\left( x\right) \} \) . Deduce Danskin’s formula \[ {f}^{\prime }\left( {x;v}\right) = \max \left\{ {\left\langle {{g}_{x}\left( {x, u}\right), v}\right\rangle : u \in M\left( x\right) }\right\} . \] 9.14. Let \( X \) be a Banach space, and set \( f\left( x\right) \mathrel{\text{:=}} \parallel x\parallel \) . If \( \zeta \in {\partial }_{C}f\left( x\right) \) for some \( x \neq 0 \), then \( \parallel \zeta {\parallel }_{ * } = 1 \) . 9.15. Let \( f : {\mathbb{R}}^{m} \times {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be locally Lipschitz, and such that for each \( x \in {\mathbb{R}}^{m} \), the function \( y \mapsto f\left( {x, y}\right) \) is convex. Then if \( \left( {\theta ,\zeta }\right) \in {\partial }_{C}f\left( {x, y}\right) \) , it follows that \( \zeta \in {\partial }_{C}f\left( {x, \cdot }\right) \left( y\right) \) . (Problem 9.4 shows that this is false in general if the partial convexity hypothesis is deleted.) 9.16. We define a continuous function \( f : \left\lbrack {-1,1}\right\rbrack \rightarrow \mathbb{R} \) as follows: (i) \( f\left( x\right) = 0 \) when \( x = \pm {2}^{-n}, n = 0,1,2,\ldots \) . (ii) \( f\left( x\right) = x \) when \( x = \pm 3/{2}^{2n}, n = 0,1,2,3,\ldots \) . (iii) \( f\left( x\right) = - x \) when \( x = \pm 3/{2}^{{2n} + 1}, n = 1,2,3,\ldots \) . (iv) Between any two successive values of \( x \) cited above, \( f \) is affine. (a) Prove that \( f \) is Lipschitz, and that the directional derivatives \( {f}^{\prime }\left( {0;1}\right) \) and \( {f}^{\prime }\left( {0; - 1}\right) \) fail to exist. (b) Show that the subderivate \( {Df}\left( {0;1}\right) \) (see Problem 9.8) equals -1 . (c) Show that \( {\partial }_{P}f\left( 0\right) = \varnothing \) and \( {\partial }_{L}f\left( 0\right) = {\partial }_{C}f\left( 0\right) = \left\lbrack {-3,3}\right\rbrack \) . 9.17. Let \( A, C, D \) be convex \( {w}^{ * } \) -closed bounded subsets of \( {X}^{ * } \) such that \( A + C \subset D + C \) . Prove that \( A \subset D \) . 9.18. Let \( F : X \rightarrow {\mathbb{R}}^{n} \) be locally Lipschitz, let \( A \subset {\mathbb{R}}^{n} \) be compact. We set \[ \theta \left( x\right) \mathrel{\text{:=}} \mathop{\max }\limits_{{\alpha \in A}}\langle \alpha, F\left( x\right) \rangle . \] (a) Show that \( \theta \) is locally Lipschitz. (b) If the maximum defining \( \theta \left( x\right) \) is attained at a single \( \alpha \in A \), prove that \[ \partial \theta \left( x\right) \subset \partial \langle \alpha, F\left( \cdot \right) \rangle \left( x\right) . \] 9.19. Let \( f,{g}_{i},{h}_{j} \) be locally Lipschitz functions on a Hilbert space \( X \) , \( i \in I \mathrel{\text{:=}} \{ 1,2,\ldots, m\}, j \in J \mathrel{\text{:=}} \{ 1,2,\ldots, n\} \), and let \( C \) be a closed subset of \( X \) . We assume that \( \bar{x} \) is a local solution of the following optimization problem: \[ \text{minimize}\left\{ {f\left( x\right) : {g}_{i}\left( x\right) \leq 0\left( {i \in I}\right) ,{h}_{j}\left( x\right) = 0\left( {j \in J}\right), x \in C}\right\} \text{.} \] We will prove the Fritz John necessary conditions, namely the following multiplier rule: There exist \( {\lambda }_{0} \in \mathbb{R},\gamma \in {\mathbb{R}}^{m},\lambda \in {\mathbb{R}}^{n} \) with \[ {\lambda }_{0} = 0\text{ or }1,\;\left( {{\lambda }_{0},\gamma ,\lambda }\right) \neq 0,\;\gamma \geq 0,\;\left\langle {\gamma, g\left( \bar{x}\right) }\right\rangle = 0, \] such that \[ 0 \in {\partial }_{C}\left\{ {{\lambda }_{0}f+\langle \gamma, g\rangle +\langle \lambda, h\rangle }\right\} \left( \bar{x}\right) + {N}_{C}^{L}\left( \bar{x}\right) . \] (a) We may assume that \( C \) is bounded, that \( f \) is globally Lipschitz and bounded below, and that \( \bar{x} \) is a global solution. We set \[ A \mathrel{\text{:=}} \left\{ {\left( {{\lambda }_{0},\gamma ,\lambda }\right) = : \alpha \in \mathbb{R} \times {\mathbb{R}}^{m} \times {\mathbb{R}}^{n} : {\lambda }_{0} \geq 0,\gamma \geq 0,\begin{Vmatrix}\left( {{\lambda }_{0},\gamma ,\lambda }\right) \end{Vmatrix}}\right\} = 1\} . \] For \( \varepsilon > 0 \) fixed, we define \[ {F}_{\varepsilon }\left( x\right) \mathrel{\text{:=}} \left( {f\left( x\right) - f\left( \bar{x}\right) + \varepsilon, g\left( x\right), h\left( x\right) }\right) ,{\theta }_{\varepsilon }\left( x\right) \mathrel{\text{:=}} \mathop{\max }\limits_{{\alpha \in A}}\left\lan
1012_(GTM178)Nonsmooth Analysis and Control Theory
36
gamma, g\left( \bar{x}\right) }\right\rangle = 0, \] such that \[ 0 \in {\partial }_{C}\left\{ {{\lambda }_{0}f+\langle \gamma, g\rangle +\langle \lambda, h\rangle }\right\} \left( \bar{x}\right) + {N}_{C}^{L}\left( \bar{x}\right) . \] (a) We may assume that \( C \) is bounded, that \( f \) is globally Lipschitz and bounded below, and that \( \bar{x} \) is a global solution. We set \[ A \mathrel{\text{:=}} \left\{ {\left( {{\lambda }_{0},\gamma ,\lambda }\right) = : \alpha \in \mathbb{R} \times {\mathbb{R}}^{m} \times {\mathbb{R}}^{n} : {\lambda }_{0} \geq 0,\gamma \geq 0,\begin{Vmatrix}\left( {{\lambda }_{0},\gamma ,\lambda }\right) \end{Vmatrix}}\right\} = 1\} . \] For \( \varepsilon > 0 \) fixed, we define \[ {F}_{\varepsilon }\left( x\right) \mathrel{\text{:=}} \left( {f\left( x\right) - f\left( \bar{x}\right) + \varepsilon, g\left( x\right), h\left( x\right) }\right) ,{\theta }_{\varepsilon }\left( x\right) \mathrel{\text{:=}} \mathop{\max }\limits_{{\alpha \in A}}\left\langle {\alpha ,{F}_{\varepsilon }\left( x\right) }\right\rangle . \] Then \( {\theta }_{\varepsilon } > 0 \) on \( C \), and \( {\theta }_{\varepsilon }\left( \bar{x}\right) = \varepsilon \) . (b) There exists \( {x}_{\varepsilon } \in C \cap B\left( {\bar{x};\sqrt{\varepsilon }}\right) \) and a constant \( K \) such that \[ 0 \in {\partial }_{L}{\theta }_{\varepsilon }\left( {x}_{\varepsilon }\right) + K{\partial }_{L}{d}_{C}\left( {x}_{\varepsilon }\right) + 4\sqrt{\varepsilon }B. \] (Hint. Invoke the Minimization Principle Theorem 1.4.2.) (c) There exists \( {\alpha }_{\varepsilon } \in A \) such that \[ 0 \in {\partial }_{C}\left\langle {{\alpha }_{\varepsilon },{F}_{\varepsilon }\left( \cdot \right) }\right\rangle \left( {x}_{\varepsilon }\right) + K{\partial }_{L}{d}_{C}\left( {x}_{\varepsilon }\right) + 4\sqrt{\varepsilon }B. \] (Hint. See the previous problem.) (d) Pass to the limit as \( \varepsilon \downarrow 0 \) . 9.20. Let \( X \) and \( Y \) be Banach spaces and \( A : X \rightarrow Y \) a continuous linear operator. Let \( \Omega \) and \( Z \) be closed convex subsets of \( X \) and \( Y \), respectively, and consider for each \( y \) in \( Y \) the problem: \[ \text{minimize}f\left( x\right) : x \in \Omega ,\;{Ax} \in Z + y\text{,} \] where \( f : X \rightarrow \mathbb{R} \) is given. Let \( V\left( y\right) \) denote the value of this problem for a given \( y \) . Now suppose that there exists a bounded subset \( K \) of \( X \) such that \( f \) is Lipschitz on \( K \), and such that for all \( y \) sufficiently near 0, we have \[ x \in \Omega ,\;{Ax} \in Z + y \Rightarrow x \in K. \] (a) If the condition \( 0 \in \operatorname{int}\{ {A\Omega } - Z\} \) is satisfied, then \( V\left( \cdot \right) \) is Lipschitz near 0. (b) If in addition \( \bar{x} \) is a solution of the problem obtained when \( y = 0 \) , then there exist \( \zeta \in \partial f\left( \bar{x}\right) \) and \( \theta \in \partial V\left( 0\right) \) such that \[ - \zeta + {A}^{ * }\theta \in {N}_{\Omega }\left( \bar{x}\right) \;\text{ and }\;\theta \in - {N}_{Z}\left( {A\bar{x}}\right) . \] (Hint. For any \( x \) in \( \Omega \) and \( z \) in \( Z \), we have \[ V\left( {{Ax} - z}\right) \leq f\left( x\right) \text{.)} \] ## 3 Special Topics Evidemment, évidemment On danse encore Sur les accords Qu'on aimait tant.... Mais pas comme avant. -Michel Berger et France Gall, Evidemment In this chapter we study a number of different issues, each of interest in its own right. All the results obtained here build upon, and in some cases complement, those of the preceding chapters, and several of them address problems discussed in the Introduction. This is the case, for example, of the first section on constrained optimization. Some of the results of §§5 and 6 will be called upon in Chapter 4. ## 1 Constrained Optimization and Value Functions Consider the problem of minimizing \( f\left( x\right) \) over those points \( x \) satisfying \( h\left( x\right) = 0 \), where \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) and \( h : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{m} \) are given (let us say smooth) functions. As we did in Chapter 0, let us associate with this single problem in constrained optimization a family of related problems \( P\left( \alpha \right) \) parametrized by \( \alpha \in {\mathbb{R}}^{m} \) : Problem \( P\left( \alpha \right) \) : minimize \( f\left( x\right) \) subject to \( h\left( x\right) + \alpha = 0 \) . The value function \( V\left( \cdot \right) \) associated with this scheme is the function whose value at \( \alpha \) is the minimum (or infimum) in the problem \( P\left( \alpha \right) \) : \[ V\left( \alpha \right) \mathrel{\text{:=}} \inf \{ f\left( x\right) : h\left( x\right) + \alpha = 0\} . \] In general, \( V \) will take values in \( \left\lbrack {-\infty ,\infty }\right\rbrack \), the value \( + \infty \) corresponding to cases in which the feasible set for \( P\left( \alpha \right) \), i.e., the set \[ \Phi \left( \alpha \right) \mathrel{\text{:=}} \left\{ {x \in {\mathbb{R}}^{n} : h\left( x\right) + \alpha = 0}\right\} \] is empty. As we pointed out in Chapter 0, the very definition of \( V \) implies that for any \( x \in {\mathbb{R}}^{n} \) we have \( f\left( x\right) \geq V\left( {-h\left( x\right) }\right) \) . If \( \sum \left( \alpha \right) \) denotes the (possibly empty) set of solutions to \( P\left( \alpha \right) \) (i.e., those \( x \in \Phi \left( \alpha \right) \) such that \( V\left( \alpha \right) = f\left( x\right) \) ), then equality holds in the preceding for any \( {x}_{0} \in \sum \left( 0\right) \) . Thus, if \( {V}^{\prime }\left( 0\right) \) exists, we derive the conclusion \[ {f}^{\prime }\left( {x}_{0}\right) + \nabla V{\left( 0\right) }^{ * }{h}^{\prime }\left( {x}_{0}\right) = 0 \] that is to say, the Lagrange Multiplier Rule. (Our convention is to identify \( {h}^{\prime }\left( {x}_{0}\right) \) with the \( m \times n \) Jacobian matrix; the gradient of \( V,\nabla V \), is an \( m \times 1 \) column vector, as an element of \( {\mathbb{R}}^{m} \) ; * denotes transpose.) As we have said, \( V \) may not be differentiable at 0, so that this line of reasoning, while attractive, is problematic. A key to a more realistic approach is to observe that the argument above can be salvaged if we possess a vector \( \zeta \) belonging to the proximal subdifferential of \( V \) at \( 0,{\partial }_{P}V\left( 0\right) \) . Let us see why. The proximal subgradient inequality (Theorem 1.2.5) asserts that for some \( \sigma \geq 0 \), for all \( \alpha \) sufficiently near 0, we have \[ V\left( \alpha \right) - V\left( 0\right) + \sigma \parallel \alpha {\parallel }^{2} \geq \langle \zeta ,\alpha \rangle \] Our given point \( {x}_{0} \in \sum \left( 0\right) \) satisfies \( f\left( {x}_{0}\right) = V\left( 0\right) \) . If \( x \) is sufficiently near \( {x}_{0} \) , then \( h\left( x\right) \) is close to \( h\left( {x}_{0}\right) = 0 \), and so the choice \( \alpha = - h\left( x\right) \) is legitimate in the preceding inequality. Finally, observe again that \( f\left( x\right) \geq V\left( {-h\left( x\right) }\right) \) . Substituting into the proximal subgradient inequality and rearranging leads to \[ f\left( x\right) + \langle \zeta, h\left( x\right) \rangle + \sigma \parallel h\left( x\right) {\parallel }^{2} \geq f\left( {x}_{0}\right) , \] an inequality which holds for all \( x \) near \( {x}_{0} \) . This amounts to saying that the function \[ x \mapsto f\left( x\right) + \langle \zeta, h\left( x\right) \rangle + \sigma \parallel h\left( x\right) {\parallel }^{2} \] admits a local minimum at \( x = {x}_{0} \), whence \[ {f}^{\prime }\left( {x}_{0}\right) + {\zeta }^{ * }{h}^{\prime }\left( {x}_{0}\right) = 0. \] We recover the Multiplier Rule in terms similar to the above, but, with \( {V}^{\prime }\left( 0\right) \) replaced by any element \( \zeta \) of \( {\partial }_{P}V\left( 0\right) \) . An important objection persists, however, for there is no guarantee that \( {\partial }_{P}V\left( 0\right) \) is nonempty. The theorem below deals with this contingency by invoking the Proximal Density Theorem to find points \( {\alpha }_{i} \) near 0 at which \( {\partial }_{P}V\left( {\alpha }_{i}\right) \neq \varnothing \), and passing to the limit as \( {\alpha }_{i} \rightarrow 0 \) . This approach requires only that \( V\left( \cdot \right) \) be lower semicontinuous, which (unlike differentiability!) is easy to guarantee a priori. For this purpose we introduce the following growth hypothesis: 1.1. Growth Hypothesis. For every \( r, s \in \mathbb{R} \), the following set is bounded: \[ \left\{ {x \in {\mathbb{R}}^{n} : f\left( x\right) \leq r,\parallel h\left( x\right) \parallel \leq s}\right\} . \] 1.2. Exercise. Under Growth Hypothesis 1.1, prove the following facts: (a) \( V\left( \alpha \right) < \infty \) precisely when \( \Phi \left( \alpha \right) \neq \varnothing \), and in that case \( \sum \left( \alpha \right) \) is nonempty. (b) \( V : {\mathbb{R}}^{m} \rightarrow ( - \infty ,\infty \rbrack \) is lower semicontinuous, and so \( V \in \mathcal{F}\left( {\mathbb{R}}^{m}\right) \) if \( V \) is somewhere finite. We will see presently how the following theorem is the generator of necessary conditions and solvability results. 1.3. Theorem. Let \( f \) and \( h \) be \( {C}^{1} \), let Growth Hypothesis 1.1 hold, and suppose \( V\left( 0\right) < + \infty \) . Then there exists a sequence \( \left\{ {\alpha }_{i}\right\} \) converging to 0, with \( V\left( {\alpha }_{i}\right) \rightarrow V\left( 0\right) \), and points \( {\zeta }_{i} \in {\partial }_{P}V\left( {\alpha }_{i}\right) ,{x}_{i} \in \sum \left( {\alpha }_{i}\right) \), such that \[ {f}^{\prime }\left( {x}_{i}\right) + {\zeta }_{i}^{ * }{h}^{\prime }\left( {x}_{i}\right) = 0,\;i = 1,2,\ldots \] Proof. We use Theorem 1.3.1 to deduce the existence of a sequence \( \left
1012_(GTM178)Nonsmooth Analysis and Control Theory
37
\varnothing \), and in that case \( \sum \left( \alpha \right) \) is nonempty. (b) \( V : {\mathbb{R}}^{m} \rightarrow ( - \infty ,\infty \rbrack \) is lower semicontinuous, and so \( V \in \mathcal{F}\left( {\mathbb{R}}^{m}\right) \) if \( V \) is somewhere finite. We will see presently how the following theorem is the generator of necessary conditions and solvability results. 1.3. Theorem. Let \( f \) and \( h \) be \( {C}^{1} \), let Growth Hypothesis 1.1 hold, and suppose \( V\left( 0\right) < + \infty \) . Then there exists a sequence \( \left\{ {\alpha }_{i}\right\} \) converging to 0, with \( V\left( {\alpha }_{i}\right) \rightarrow V\left( 0\right) \), and points \( {\zeta }_{i} \in {\partial }_{P}V\left( {\alpha }_{i}\right) ,{x}_{i} \in \sum \left( {\alpha }_{i}\right) \), such that \[ {f}^{\prime }\left( {x}_{i}\right) + {\zeta }_{i}^{ * }{h}^{\prime }\left( {x}_{i}\right) = 0,\;i = 1,2,\ldots \] Proof. We use Theorem 1.3.1 to deduce the existence of a sequence \( \left\{ {\alpha }_{i}\right\} \) such that \( {\alpha }_{i} \rightarrow 0, V\left( {\alpha }_{i}\right) \rightarrow V\left( 0\right) \), and such that \( {\partial }_{P}V\left( {\alpha }_{i}\right) \neq \varnothing \) . Pick \( {\zeta }_{i} \in \) \( {\partial }_{P}V\left( {\alpha }_{i}\right) \) and \( {x}_{i} \in \sum \left( {\alpha }_{i}\right) \) (this set is nonempty by Exercise 1.2). Then, as in the argument presented above, the proximal subgradient inequality for \( V \) at \( {\alpha }_{i} \) translates to the condition \[ f\left( x\right) + \left\langle {{\zeta }_{i}, h\left( x\right) }\right\rangle + {\sigma }_{i}{\begin{Vmatrix}h\left( x\right) - h\left( {x}_{i}\right) \end{Vmatrix}}^{2} \geq f\left( {x}_{i}\right) + \left\langle {{\zeta }_{i},{\alpha }_{i}}\right\rangle \] for all \( x \) near \( {x}_{i} \) . Setting the derivative equal to 0 gives the final conclusion of the theorem. The question now becomes that of passing to the limit in the context of the sequences provided by the theorem. How we proceed will be conditioned by the specific goal we have in mind. Suppose to begin with that we require information on the sensitivity of the problem \( P\left( 0\right) \) with respect to perturbations of the equality constraint. This might arise, for example, from a numerical algorithm which, applied to \( P\left( 0\right) \), can be shown to yield a solution not to \( P\left( 0\right) \) itself, but to a perturbed problem \( P\left( \alpha \right) \) . The question then is: How different is \( V\left( \alpha \right) \) from \( V\left( 0\right) \) ? Differential information about \( V \) at 0 becomes relevant. To state a result along these lines, we introduce the multiplier set \( M\left( x\right) \) corresponding to \( x \) : \[ M\left( x\right) \mathrel{\text{:=}} \left\{ {\zeta \in {\mathbb{R}}^{m} : {f}^{\prime }\left( x\right) + {\zeta }^{ * }{h}^{\prime }\left( x\right) = 0}\right\} . \] The hypotheses of Theorem 1.3 remain in force. 1.4. Corollary. Suppose that for every \( x \in \sum \left( 0\right) \), the Jacobian \( {h}^{\prime }\left( x\right) \) is of maximal rank. Then \( V\left( \cdot \right) \) is Lipschitz near 0, and we have \[ \varnothing \neq {\partial }_{L}V\left( 0\right) \subset \mathop{\bigcup }\limits_{{x \in \sum \left( 0\right) }}M\left( x\right) \] Proof. We consider the sequences provided by the theorem, and deduce from the conditions \( f\left( {x}_{i}\right) \rightarrow V\left( 0\right), h\left( {x}_{i}\right) = {\alpha }_{i} \rightarrow 0 \) that, in light of Growth Hypothesis 1.1, the sequence \( \left\{ {x}_{i}\right\} \) is bounded. We extract a subsequence converging to a limit \( {x}_{0} \), without relabeling. It follows from the continuity of \( f \) and \( h \) that \( f\left( {x}_{0}\right) = V\left( 0\right), h\left( {x}_{0}\right) = 0 \), whence \( {x}_{0} \in \sum \left( 0\right) \) . Now extract a further subsequence to arrange either \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \rightarrow \infty \) or else \( {\zeta }_{i} \rightarrow {\zeta }_{0} \in {\mathbb{R}}^{m} \) . Let us immediately dispose of the first possibility: Divide across by \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \) in the equation \[ {f}^{\prime }\left( {x}_{i}\right) + {\zeta }_{i}^{ * }{h}^{\prime }\left( {x}_{i}\right) = 0 \] take a further subsequence to arrange to have \( {\zeta }_{i}/\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \) converge to a (necessarily nonzero) limit \( \lambda \in {\mathbb{R}}^{m} \), and arrive at \( {\lambda }^{ * }{h}^{\prime }\left( {x}_{0}\right) = 0 \), contradicting the maximal rank hypothesis. Thus \( \left\{ {\zeta }_{i}\right\} \) is necessarily bounded; the arbitrariness of this sequence of proximal subgradients (having \( V\left( {\alpha }_{i}\right) \rightarrow V\left( 0\right) \) ) implies that for some \( \varepsilon > 0 \), for some \( K \), for all \( \alpha \in B\left( {0,\varepsilon }\right) \) for which \( \left| {V\left( \alpha \right) - V\left( 0\right) }\right| < \varepsilon \) and for all \( \zeta \in {\partial }_{P}V\left( \alpha \right) \), we have \( \parallel \zeta \parallel \leq K \) . This is a proximal criterion guaranteeing that \( V \) is Lipschitz on a neighborhood of 0 (Problem 1.11.11), as stated. To conclude the proof, observe that any \( {\zeta }_{0} \in {\partial }_{L}V\left( 0\right) \) is the limit of a sequence \( {\zeta }_{i} \) as above, and that we have, along the sequence, \[ {f}^{\prime }\left( {x}_{i}\right) + {\zeta }_{i}^{ * }{h}^{\prime }\left( {x}_{i}\right) = 0. \] It suffices to pass to the limit to see that \( {\zeta }_{0} \in M\left( {x}_{0}\right) \) . Remark. We should not expect an exact formula for \( {\partial }_{L}V \) in terms of multipliers, since there is in general a real distinction between actual minimization (as measured by \( V \) ) and stationarity conditions (provided by multipliers). In the "fully convex" case of the problem the distinction vanishes, and the formula becomes exact; see the end-of-chapter problems. Given a particular solution \( {x}_{0} \) to \( P\left( 0\right) \), it is not clear at first glance that Corollary 1.4 will imply the Multiplier Rule, namely that \( M\left( {x}_{0}\right) \neq \varnothing \) . Yet it does, as we now see. ## 1.5. Exercise. (a) Let \( {x}_{0} \) solve \( P\left( 0\right) \), where \( {h}^{\prime }\left( {x}_{0}\right) \) has maximal rank. Prove that \( M\left( {x}_{0}\right) \neq \varnothing \) . (Hint. Modify the problem by adding \( {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2} \) to \( f \), so that \( {x}_{0} \) becomes the unique solution to \( P\left( 0\right) \) ; How is \( M\left( {x}_{0}\right) \) modified?) (b) Show that the rank hypothesis cannot be deleted by considering \( n = m = 1, f\left( x\right) = x, h\left( x\right) = {x}^{2} \) . Another issue that comes to mind is that of relaxing the smoothness hypothesis on \( f \) and \( h \) . An analysis of the argument used to prove Corollary 1.4 shows that an extension to nonsmooth data would hinge upon being able to pass to the limit in the relation \[ 0 \in {\partial }_{P}\left\{ {f\left( \cdot \right) + \left\langle {{\zeta }_{i}, h\left( \cdot \right) }\right\rangle }\right\} \left( {x}_{i}\right) . \] Here is the required fact. 1.6. Exercise. Let \( \theta : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{k} \) be locally Lipschitz, and let \( \left\{ {\lambda }_{i}\right\} \) in \( {\mathbb{R}}^{k} \) and \( \left\{ {x}_{i}\right\} \) in \( {\mathbb{R}}^{n} \) be sequences converging to \( \lambda \) and \( x \), respectively. Suppose that \( 0 \in {\partial }_{L}\left\{ {\left\langle {{\lambda }_{i},\theta }\right\rangle \left( \cdot \right) }\right\} \left( {x}_{i}\right) \) for each \( i \) . Prove that \( 0 \in {\partial }_{L}\{ \langle \lambda ,\theta \rangle \left( \cdot \right) \rangle \left( x\right) \) . (Hint. Proposition 1.10.1.) We now redefine the multiplier set \( M\left( x\right) \) for nonsmooth data as the set of those \( \zeta \in {\mathbb{R}}^{m} \) for which \( 0 \in {\partial }_{L}\{ f\left( \cdot \right) + \langle \zeta, h\left( \cdot \right) \rangle \} \left( x\right) \) . We say that \( x \) is normal provided that \( 0 \in {\partial }_{L}\langle \zeta, h\left( \cdot \right) \rangle \left( x\right) \) implies \( \zeta = 0 \) (note that this is equivalent to the rank of \( {h}^{\prime }\left( x\right) \) being maximal if \( h \) is \( {C}^{1} \) ). Otherwise we call \( x \) abnormal. Then, upon noting that Exercise 1.2 requires only continuity of \( f \) and \( h \), and calling upon Exercise 1.6 at a certain point, our sequential argument yields the following analogue of Corollary 1.4, whose detailed proof is left as an exercise. 1.7. Theorem. Let \( f \) and \( h \) be locally Lipschitz, and let Growth Hypothesis 1.1 hold. Suppose that \( V\left( 0\right) < \infty \), and that every \( x \in \sum \left( 0\right) \) is normal. Then \( V\left( \cdot \right) \) is Lipschitz near 0, and \[ \varnothing \neq {\partial }_{L}V\left( 0\right) \subset \mathop{\bigcup }\limits_{{x \in \sum \left( 0\right) }}M\left( x\right) \] This can be used just as in the smooth case (i.e., in Exercise 1.5(a)) to deduce the multiplier rule for a given solution \( {x}_{0} \) in the case of locally Lipschitz data. The result is often stated in an equivalent form without the normality hypothesis, but with a conclusion that holds in one of two alternative forms, abnormal or normal, as in the following: 1.8. Exercise. Let \( {x}_{0} \) solve \( P\left( 0\right) \), where \( f \) and \( h \) are locally Lipschitz and Growth Hypothesis 1.1 holds. Then there exists \( \left( {{\lambda }_{0},\zeta }\right) \in \) \( \mathbb{R} \times {\mathbb{R}}^{m} \) with \( \left( {{\lambda }_{0},\zeta }\right) \neq \left( {0,0}\right) \) and \( {\lambda }_{0} = 0 \) or 1, such that \[ 0 \in {\partial }_{L}\left\{ {{\lambda }_{0}f\left( \cdot \right) +\langle \zeta, h\left( \cdot \right) \rangle }\rig
1012_(GTM178)Nonsmooth Analysis and Control Theory
38
and \[ \varnothing \neq {\partial }_{L}V\left( 0\right) \subset \mathop{\bigcup }\limits_{{x \in \sum \left( 0\right) }}M\left( x\right) \] This can be used just as in the smooth case (i.e., in Exercise 1.5(a)) to deduce the multiplier rule for a given solution \( {x}_{0} \) in the case of locally Lipschitz data. The result is often stated in an equivalent form without the normality hypothesis, but with a conclusion that holds in one of two alternative forms, abnormal or normal, as in the following: 1.8. Exercise. Let \( {x}_{0} \) solve \( P\left( 0\right) \), where \( f \) and \( h \) are locally Lipschitz and Growth Hypothesis 1.1 holds. Then there exists \( \left( {{\lambda }_{0},\zeta }\right) \in \) \( \mathbb{R} \times {\mathbb{R}}^{m} \) with \( \left( {{\lambda }_{0},\zeta }\right) \neq \left( {0,0}\right) \) and \( {\lambda }_{0} = 0 \) or 1, such that \[ 0 \in {\partial }_{L}\left\{ {{\lambda }_{0}f\left( \cdot \right) +\langle \zeta, h\left( \cdot \right) \rangle }\right\} \left( {x}_{0}\right) . \] A result such as Theorem 1.7 incorporates solvability information: Since \( V\left( \cdot \right) \) is Lipschitz near 0, therefore \( V\left( \alpha \right) \) is finite for \( \alpha \) near 0, and so the feasible set \( \Phi \left( \alpha \right) \) is nonempty for \( \alpha \) near 0 . That is to say, for \( \alpha \) near 0 the equation \( h\left( x\right) = - \alpha \) admits at least one solution. The next theorem improves on this somewhat. Notice that optimization plays no role in its formulation; however, the proof is in essence an appeal to Theorem 1.7. 1.9. Theorem. Let \( f \) and \( h \) be locally Lipschitz. Suppose that \( {x}_{0} \) is a normal point satisfying \( h\left( {x}_{0}\right) = 0 \) . Then for some \( K \), for all \( \alpha \) sufficiently near 0, the equation \( h\left( x\right) + \alpha = 0 \) admits a solution \( {x}_{\alpha } \) satisfying \( \begin{Vmatrix}{{x}_{\alpha } - {x}_{0}}\end{Vmatrix} \leq K\parallel \alpha \parallel \) Proof. Consider the problem \( P\left( \alpha \right) \) of minimizing \( \begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix} \) subject to \( h\left( x\right) + \) \( \alpha = 0 \) . Clearly \( {x}_{0} \) is the unique solution of \( P\left( 0\right) \), and Growth Hypothesis 1.1 is satisfied. Apply Theorem 1.7, and let \( K \) be a Lipschitz constant for \( V\left( \cdot \right) \) in a neighborhood of 0, say \( B\left( {0;\varepsilon }\right) \) . Then, if \( \parallel \alpha \parallel < \varepsilon \) and if \( {x}_{\alpha } \) is a solution to \( P\left( \alpha \right) \), we have \( h\left( {x}_{\alpha }\right) + \alpha = 0 \) and \[ \begin{Vmatrix}{{x}_{\alpha } - {x}_{0}}\end{Vmatrix} = V\left( \alpha \right) \leq V\left( 0\right) + K\parallel \alpha \parallel = K\parallel \alpha \parallel \] as required. Some remarks on this result are in order. First, note than when \( h \) is smooth, it reduces to a classical corollary of the Implicit Function Theorem: If the Jacobian matrix \( {h}^{\prime }\left( {x}_{0}\right) \) is of maximal rank, then \( h\left( x\right) + \alpha = 0 \) is locally solvable. We will address the issue of constructing implicit and inverse functions in \( §3 \), in a more general context, as well as how to quantify precisely the parameters of such a result. Next, observe that even when \( h \) is smooth, the proof of the theorem involves the nonsmooth cost function \( f\left( x\right) = \begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix} \) . A smooth choice such as \( {\begin{Vmatrix}x - {x}_{0}\end{Vmatrix}}^{2} \) would give only a weaker (Hölder) bound \( \begin{Vmatrix}{{x}_{\alpha } - {x}_{0}}\end{Vmatrix} \leq K\parallel \alpha {\parallel }^{1/2} \) . Finally, we remark that \( §3 \) will give a criterion in terms of the generalized Jacobian of \( h \) by which the normality assumption can be verified more directly. ## Inequality Constraints Besides equality constraints, many optimization problems feature inequality constraints. Let us consider now the problem of minimizing \( f\left( x\right) \) subject to \( g\left( x\right) \leq 0 \), where \( g : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is given and is locally Lipschitz, along with \( f \) . Much as before, we define \( P\left( \beta \right) \) to be the perturbed problem having constraint \( g\left( x\right) + \beta \leq 0 \), and associate to \( P\left( \beta \right) \) the feasible set \( \Phi \left( \beta \right) \), the solution set \( \sum \left( \beta \right) \), and the value function \( V\left( \beta \right) \) . The appropriate Growth Hypothesis is that the set \( \{ x : f\left( x\right) \leq r, g\left( x\right) \leq s\} \) be bounded for each \( r \) and \( s \in \mathbb{R} \) . The first real difference with the equality case arises when we analyze the proximal subgradient inequality for \( V \) . Let \( \gamma \in {\partial }_{P}V\left( 0\right) \), for example, and let \( {x}_{0} \in \sum \left( 0\right) \) . Then, as before, \[ V\left( \beta \right) - V\left( 0\right) + \sigma {\left| \beta \right| }^{2} \geq \langle \gamma ,\beta \rangle = {\gamma \beta } \] for all \( \beta \) near 0 . Note that \( V\left( \beta \right) \leq V\left( 0\right) \) for \( \beta \leq 0 \) (the function \( V \) is nondecreasing by its very nature), so that \( \langle \gamma ,\beta \rangle \leq \sigma {\left| \beta \right| }^{2} \) for all small \( \beta \leq 0 \) . This implies that \( \gamma \geq 0 \) . Suppose now that \( g\left( {x}_{0}\right) \) is strictly less than 0 . It follows that \( V\left( \beta \right) = V\left( 0\right) \) for any \( \beta \in \left\lbrack {g\left( {x}_{0}\right) ,0}\right\rbrack \) (why?), and the proximal subgradient inequality yields \( \langle \gamma ,\beta \rangle \leq \sigma {\left| \beta \right| }^{2} \) for all \( \beta \) near 0, so that \( \gamma = 0 \) . To summarize thus far, we have proved that at least one of \( \gamma \) or \( g\left( {x}_{0}\right) \) must be zero, a conclusion which can also be stated in the form \( {\gamma g}\left( {x}_{0}\right) = 0 \) . If \( g\left( {x}_{0}\right) < 0 \), then \( {x}_{0} \) is a local minimum for \( f \), and \( 0 \in {\partial }_{L}f\left( {x}_{0}\right) \) . Otherwise, we proceed to substitute \( \beta = - g\left( x\right) \) into the proximal subgradient inequality, for \( x \) near \( {x}_{0} \) . We deduce \[ f\left( x\right) + \sigma {\left| g\left( x\right) \right| }^{2} + {\gamma g}\left( x\right) \geq f\left( {x}_{0}\right) , \] with equality when \( x = {x}_{0} \) . This implies \( 0 \in {\partial }_{L}\{ f\left( \cdot \right) + {\gamma g}\left( \cdot \right) \} \left( {x}_{0}\right) \), a condition that actually holds in either of the two cases, since in the first case we have \( \gamma = 0 \) . These observations lead us to define the new multiplier set \( M\left( x\right) \) for the present context as those scalars \( \gamma \geq 0 \) for which \( {\gamma g}\left( x\right) = 0 \) and \( 0 \in \) \( {\partial }_{L}\{ f\left( \cdot \right) + {\gamma g}\left( \cdot \right) \} \left( x\right) \) . A point \( x \) satisfying \( g\left( x\right) \leq 0 \) is now called normal if \[ \gamma \geq 0,\;{\gamma g}\left( x\right) = 0,\;0 \in {\partial }_{L}\{ {\gamma g}\left( \cdot \right) \} \left( x\right) \Rightarrow \gamma = 0. \] With these modified definitions, the following result is visually identical to Theorem 1.7. 1.10. Exercise. With \( f, g \) as above, suppose that \( V\left( 0\right) < \infty \), and that every \( x \in \sum \left( 0\right) \) is normal. Then \( V\left( \cdot \right) \) is Lipschitz near 0, and \[ \varnothing \neq {\partial }_{L}V\left( 0\right) \subset \mathop{\bigcup }\limits_{{x \in \sum \left( 0\right) }}M\left( x\right) \] Here are the counterparts of the necessary conditions (Exercise 1.8) and the solvability consequences (Theorem 1.9) for the inequality case being considered. 1.11. Exercise. Let \( {x}_{0} \) solve \( P\left( 0\right) \), where \( f \) and \( g \) are locally Lipschitz and the (modified) Growth Hypothesis holds. Then there exist \( \left( {{\lambda }_{0},\gamma }\right) \) not both 0, with \( {\lambda }_{0} = 0 \) or 1 and \( \gamma \geq 0 \), such that \[ {\gamma g}\left( {x}_{0}\right) = 0\;\text{ and }\;0 \in {\partial }_{L}\left\{ {{\lambda }_{0}f\left( \cdot \right) + {\gamma g}\left( \cdot \right) }\right\} \left( {x}_{0}\right) . \] 1.12. Exercise. Let \( f \) and \( g \) be locally Lipschitz. Suppose that \( {x}_{0} \) is a normal point satisfying \( g\left( {x}_{0}\right) \leq 0 \) . Then for some \( K \), for all \( \beta \) sufficiently near 0, the inequality \( g\left( x\right) + \beta \leq 0 \) admits a solution \( {x}_{\beta } \) satisfying \( \begin{Vmatrix}{{x}_{\beta } - {x}_{0}}\end{Vmatrix} \leq K\max \{ 0,\beta \} \) . 1.13. Exercise. Show that for \( n = 1,{x}_{0} = 0 \), the function \( h\left( x\right) \mathrel{\text{:=}} \) \( - \left| x\right| \) is not normal in the sense of the word that prevails in Theorem 1.9, but that the same function \( g\left( x\right) \mathrel{\text{:=}} - \left| x\right| \) is normal as the term is employed in Exercise 1.12. Note that the conclusion of Exercise 1.12 holds while that of Theorem 1.9 fails. The general case of mixed equality and inequality constraints is that of minimizing \( f\left( x\right) \) over the points \( x \in {\mathbb{R}}^{n} \) satisfying \( h\left( x\right) = 0, g\left( x\right) \leq 0 \) , where \( h : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{m} \) and \( g : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{p} \) are given functions and where a vector inequality such as \( g\left( x\right) \leq 0 \) is understood component-wise. We will assume that all these functions are locally Lipschitz, and that the set \[ \left\{ {x \in {\mathbb{R}}^{n} : f\left( x\right) \leq r, g\left( x\right) \leq s,\parallel h\left( x\right) \parallel \leq t}\right\} \] is bounded for each \( r, t \) in \( \mathbb{R} \), and \( s \) in \( {\mathbb{R}}^{p
1012_(GTM178)Nonsmooth Analysis and Control Theory
39
heorem 1.9, but that the same function \( g\left( x\right) \mathrel{\text{:=}} - \left| x\right| \) is normal as the term is employed in Exercise 1.12. Note that the conclusion of Exercise 1.12 holds while that of Theorem 1.9 fails. The general case of mixed equality and inequality constraints is that of minimizing \( f\left( x\right) \) over the points \( x \in {\mathbb{R}}^{n} \) satisfying \( h\left( x\right) = 0, g\left( x\right) \leq 0 \) , where \( h : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{m} \) and \( g : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{p} \) are given functions and where a vector inequality such as \( g\left( x\right) \leq 0 \) is understood component-wise. We will assume that all these functions are locally Lipschitz, and that the set \[ \left\{ {x \in {\mathbb{R}}^{n} : f\left( x\right) \leq r, g\left( x\right) \leq s,\parallel h\left( x\right) \parallel \leq t}\right\} \] is bounded for each \( r, t \) in \( \mathbb{R} \), and \( s \) in \( {\mathbb{R}}^{p} \) . Let \( P\left( {\alpha ,\beta }\right) \) denote the problem of minimizing \( f\left( x\right) \) subject to \( h\left( x\right) + \alpha = 0, g\left( x\right) + \beta \leq 0 \), and let \( V\left( {\alpha ,\beta }\right) \) be the corresponding value function. The Lagrangian of the problem is the function \( \mathcal{L} : {\mathbb{R}}^{n} \times {\mathbb{R}}^{p} \times {\mathbb{R}}^{m} \rightarrow \mathbb{R} \) defined by \[ \mathcal{L}\left( {x,\gamma ,\zeta }\right) \mathrel{\text{:=}} f\left( x\right) + \langle \gamma, g\left( x\right) \rangle + \langle \zeta, h\left( x\right) \rangle . \] For any point \( x \) feasible for \( P\left( {0,0}\right) \), the set \( M\left( x\right) \) of multipliers for \( x \) consists of those \( \left( {\gamma ,\zeta }\right) \in {\mathbb{R}}^{p} \times {\mathbb{R}}^{m} \) satisfying \[ \gamma \geq 0,\;\langle \gamma, g\left( x\right) \rangle = 0,\;0 \in {\partial }_{L}\mathcal{L}\left( {\cdot ,\gamma ,\zeta }\right) \left( x\right) . \] The point \( x \) is now called normal if \[ \gamma \geq 0,\;\langle \gamma, g\left( x\right) \rangle = 0 \] \[ 0 \in {\partial }_{L}\{ \langle \gamma, g\left( \cdot \right) \rangle + \langle \zeta, h\left( \cdot \right) \rangle \} \left( x\right) \Rightarrow \gamma = 0,\zeta = 0, \] and otherwise is called abnormal. We suppose that \( V\left( {0,0}\right) < \infty \) . In light of the preceding, the genesis of the following theorem should be clear. Its detailed proof is a substantial exercise. ## 1.14. Theorem. (a) If every \( x \in \sum \left( {0,0}\right) \) is normal, then \( V \) is Lipschitz near \( \left( {0,0}\right) \) and we have \[ \varnothing \neq {\partial }_{L}V\left( {0,0}\right) \subset \mathop{\bigcup }\limits_{{x \in \sum \left( {0,0}\right) }}M\left( x\right) \] (b) If \( x \) is any solution to \( P\left( {0,0}\right) \), then either \( x \) is abnormal or else \( M\left( x\right) \neq \varnothing \) . 1.15. Corollary. Let \( {x}_{0} \) be a point satisfying \( h\left( {x}_{0}\right) = 0, g\left( {x}_{0}\right) \leq 0 \), and suppose that \( {x}_{0} \) is normal. Then for some \( K > 0 \), for all \( \alpha \) and \( \beta \) near 0, there exists a point \( {x}_{\alpha ,\beta } \) such that \[ h\left( {x}_{\alpha ,\beta }\right) = 0,\;g\left( {x}_{\alpha ,\beta }\right) \leq 0,\;\begin{Vmatrix}{{x}_{\alpha ,\beta } - {x}_{0}}\end{Vmatrix} \leq K\left( {\parallel \alpha \parallel + \mathop{\sum }\limits_{{i = 1}}^{p}\max \left\{ {0,{\beta }_{i}}\right\} }\right) . \] We remark that the value function approach illustrated above has proven to be a robust and effective line of attack for many other more complex optimization problems than the finite-dimensional ones considered here, including a variety of problems in control theory. ## 2 The Mean Value Inequality In this section we will prove a "multidirectional" Mean Value Theorem, a result that is novel even in the context of smooth multivariable calculus, and one that is a cornerstone of nonsmooth analysis. It is instructive to begin by recalling the statement of the classical Mean Value Theorem, which plays a fundamental role in analysis by relating the values of a function to its derivative at an intermediate point. For the purpose of comparisons to come, it is also worth noting the optimization-based proof of this familiar result. Recall that given two points \( x \) and \( y \in X \), the open line segment connecting them is denoted \( \left( {x, y}\right) \) : \[ \left( {x, y}\right) \mathrel{\text{:=}} \{ {tx} + \left( {1 - t}\right) y : 0 < t < 1\} . \] If \( t = 0 \) and \( t = 1 \) are allowed, we obtain the closed line segment \( \left\lbrack {x, y}\right\rbrack \) . Throughout this section, \( X \) is a Hilbert space. 2.1. Proposition. Suppose \( x, y \in X \) and that \( f \in \mathcal{F} \) is Gâteaux differentiable on a neighborhood of \( \left\lbrack {x, y}\right\rbrack \) . Then there exists \( z \in \left( {x, y}\right) \) such that \[ f\left( y\right) - f\left( x\right) = \left\langle {{f}_{G}^{\prime }\left( z\right), y - x}\right\rangle . \] (1) Proof. Define \( \dot{g} : \left\lbrack {0,1}\right\rbrack \rightarrow \mathbb{R} \) by \[ g\left( t\right) \mathrel{\text{:=}} f\left( {{ty} + \left( {1 - t}\right) x}\right) - {tf}\left( y\right) - \left( {1 - t}\right) f\left( x\right) . \] Then \( g \) is continuous on \( \left\lbrack {0,1}\right\rbrack \), differentiable on \( \left( {0,1}\right) \), and satisfies \( g\left( 0\right) = \) \( g\left( 1\right) = 0 \) . It follows that there exists a point \( \bar{t} \in \left( {0,1}\right) \) that is either a maximum or a minimum of \( g \) on \( \left\lbrack {0,1}\right\rbrack \), and thus satisfies \( {g}^{\prime }\left( \bar{t}\right) = 0 \) . We calculate \( {g}^{\prime }\left( \bar{t}\right) \) by referring directly to the definition of the Gâteaux derivative at \( z \mathrel{\text{:=}} \bar{t}y + \left( {1 - \bar{t}}\right) x = x + \bar{t}\left( {y - x}\right) \), and conclude that \[ 0 = {g}^{\prime }\left( \bar{t}\right) = \left\langle {{f}_{G}^{\prime }\left( z\right) ,\left( {y - x}\right) }\right\rangle - \left( {f\left( y\right) - f\left( x\right) }\right) . \] Hence (1) holds. A typical application of the Mean Value Theorem is to deduce monotonicity. For example: suppose that \( \left\langle {{f}_{G}^{\prime }\left( z\right), y - x}\right\rangle \leq 0 \) for all \( z \in \left( {x, y}\right) \) . Then it follows immediately that \( f\left( y\right) \leq f\left( x\right) \) . In this, and in fact in most applications, it suffices to invoke the Mean Value Theorem in its inequality form; i.e., the form in which (1) is replaced by \[ f\left( y\right) - f\left( x\right) \leq \left\langle {{f}_{G}^{\prime }\left( z\right), y - x}\right\rangle \] When \( f \) is nondifferentiable, the inequality form provides the appropriate model for generalization. To illustrate, consider the function \( f\left( x\right) \mathrel{\text{:=}} - \left| x\right| \) on \( \mathbb{R} \) . There exists \( z \in \left( {-1,1}\right) \) and \( \zeta \in {\partial }_{P}f\left( z\right) \) such that \[ 0 = f\left( 1\right) - f\left( {-1}\right) \leq \langle \zeta ,1 - \left( {-1}\right) \rangle = {2\zeta } \] but equality (i.e., \( \zeta = 0 \) ) never holds. In a nonsmooth setting, it turns out as well that some tolerance must be present in the resulting inequality. For example, if \( f\left( {{x}_{1},{x}_{2}}\right) \mathrel{\text{:=}} - \left| {x}_{2}\right| \), then \( f\left( {-1,0}\right) = f\left( {1,0}\right) \), but the difference (0) between these two values of \( f \) cannot be estimated by \( {\partial }_{P}f\left( z\right) \) for any \( z \) on the line segment between \( \left( {-1,0}\right) \) and \( \left( {1,0}\right) \), since \( {\partial }_{P}f\left( z\right) \) is empty for all such \( z \) . We will need to allow \( z \) to stray a bit from the line segment. Besides nondifferentiability, the Mean Value Inequality that we will present possesses a novel, perhaps surprising feature: multidirectionality; i.e., an estimate that is maintained uniformly over many directions. This new aspect greatly enhances the range of applications of the result, even in a smooth setting. Let us describe it in a particular two-dimensional case, in which we seek to compare the value of a Gâteaux differentiable continuous function \( f \) at \( \left( {0,0}\right) \) to the set of its values on the line segment \[ Y \mathrel{\text{:=}} \{ \left( {1, t}\right) : 0 \leq t \leq 1\} . \] For each \( t \in \left\lbrack {0,1}\right\rbrack \), the usual Mean Value Theorem applied to the two points \( \left( {0,0}\right) \) and \( {y}_{t} \mathrel{\text{:=}} \left( {1, t}\right) \) asserts the existence of a point \( {z}_{t} \) lying in the open segment connecting \( \left( {0,0}\right) \) and \( {y}_{t} \) such that \[ f\left( {y}_{t}\right) - f\left( 0\right) \leq \left\langle {{f}_{G}^{\prime }\left( {z}_{t}\right) ,{y}_{t}}\right\rangle \] We deduce \[ \mathop{\min }\limits_{{y \in Y}}\left( {f\left( y\right) - f\left( 0\right) }\right) \leq \left\langle {{f}_{G}^{\prime }\left( {z}_{t}\right) ,{y}_{t}}\right\rangle \] (2) The new Mean Value Inequality asserts that for some \( z \) in the triangle which is the convex hull of \( \{ 0\} \cup Y \), we have \[ \mathop{\min }\limits_{{y \in Y}}\left( {f\left( y\right) - f\left( 0\right) }\right) \leq \left\langle {{f}_{G}^{\prime }\left( z\right) ,{y}_{t}}\right\rangle \forall t \in \left\lbrack {0,1}\right\rbrack . \] (3) While these conclusions appear quite similar, there is an important difference: (3) holds for the same given \( z \), uniformly for all \( {y}_{t} \), in contrast to (2), where \( {z}_{t} \) depended on \( {y}_{t} \) . To further underline the difference, consider now the case in which \( Y \) is the closed unit ball \( \bar{B} \) in \( X \) . For a continuous Gâteaux differentiable function \( f \), the new Mean Value Inequality asserts t
1012_(GTM178)Nonsmooth Analysis and Control Theory
40
rangle \] We deduce \[ \mathop{\min }\limits_{{y \in Y}}\left( {f\left( y\right) - f\left( 0\right) }\right) \leq \left\langle {{f}_{G}^{\prime }\left( {z}_{t}\right) ,{y}_{t}}\right\rangle \] (2) The new Mean Value Inequality asserts that for some \( z \) in the triangle which is the convex hull of \( \{ 0\} \cup Y \), we have \[ \mathop{\min }\limits_{{y \in Y}}\left( {f\left( y\right) - f\left( 0\right) }\right) \leq \left\langle {{f}_{G}^{\prime }\left( z\right) ,{y}_{t}}\right\rangle \forall t \in \left\lbrack {0,1}\right\rbrack . \] (3) While these conclusions appear quite similar, there is an important difference: (3) holds for the same given \( z \), uniformly for all \( {y}_{t} \), in contrast to (2), where \( {z}_{t} \) depended on \( {y}_{t} \) . To further underline the difference, consider now the case in which \( Y \) is the closed unit ball \( \bar{B} \) in \( X \) . For a continuous Gâteaux differentiable function \( f \), the new Mean Value Inequality asserts that for some \( z \in \bar{B} \), we have \[ \mathop{\inf }\limits_{\bar{B}}f - f\left( 0\right) \leq \left\langle {{f}_{G}^{\prime }\left( z\right), y}\right\rangle \forall y \in \bar{B} \] Assume now that \( \begin{Vmatrix}{f}_{G}^{\prime }\end{Vmatrix} \geq 1 \) on \( \bar{B} \) . Then, taking \( y = - {f}_{G}^{\prime }\left( z\right) /\begin{Vmatrix}{{f}_{G}^{\prime }\left( z\right) }\end{Vmatrix} \) in the inequality leads to \[ \mathop{\inf }\limits_{\bar{B}}f - f\left( 0\right) \leq - 1 \] (this is where we exploit the uniformity of the conclusion). Applied to \( - f \) , the same reasoning yields \[ \mathop{\sup }\limits_{\bar{B}}f - f\left( 0\right) \geq 1 \] whence \[ \mathop{\sup }\limits_{\bar{B}}f - \mathop{\inf }\limits_{\bar{B}}f \geq 2 \] The reader is invited to ponder how the usual Mean Value Theorem is inadequate to yield this conclusion. Further, let us note that the failure of \( {f}^{\prime } \) to exist even at a single point can make a difference and will need to be taken account of: Witness \( f\left( x\right) \mathrel{\text{:=}} \parallel x\parallel \), for which we have \[ \mathop{\sup }\limits_{\bar{B}}f - \mathop{\inf }\limits_{\bar{B}}f = 1 \] in contrast to the above, even though \( \begin{Vmatrix}{{f}^{\prime }\left( x\right) }\end{Vmatrix} = 1 \) for all \( x \neq 0 \) . ## A Smooth Finite-Dimensional Version The general theorem that we will present features nonsmoothness, infinite dimensions, and multidirectionality. The structure of its proof can be conveyed much more easily in the absence of the first two factors. Accordingly, we will begin by proving a version of the theorem for a differentiable function on \( {\mathbb{R}}^{n} \) . A preliminary result we will require is the following. 2.2. Proposition. Suppose \( Y \subset X \) is closed and convex, that \( f \in \mathcal{F} \) attains its minimum over \( Y \) at \( \bar{y} \), and that \( f \) has directional derivatives in all directions at \( \bar{y} \) . Then \[ {f}^{\prime }\left( {\bar{y};y - \bar{y}}\right) \geq 0\forall y \in Y. \] (4) In particular, if \( f \) is Gâteaux differentiable at \( \bar{y} \), then \[ \left\langle {{f}_{G}^{\prime }\left( \bar{y}\right) ;y - \bar{y}}\right\rangle \geq 0\forall y \in Y. \] Proof. Let \( y \in Y \), and define \( g : \left\lbrack {0,1}\right\rbrack \rightarrow \mathbb{R} \) by \( g\left( t\right) = f\left( {\bar{y} + t\left( {y - \bar{y}}\right) }\right) \) . Note that \( \bar{y} + t\left( {y - \bar{y}}\right) = {ty} + \left( {1 - t}\right) \bar{y} \) belongs to \( Y \) for each \( t \in \left\lbrack {0,1}\right\rbrack \), since \( Y \) is convex. Since \( \bar{y} \) minimizes \( f \) over \( Y \), we have \( g\left( 0\right) \leq g\left( t\right) \) for all \( t \) . Thus by letting \( t \downarrow 0 \) and using the assumption that the directional derivative \( {f}^{\prime }\left( {\bar{y}, y - \bar{y}}\right) \) exists, we conclude \[ 0 \leq \mathop{\lim }\limits_{{t \downarrow 0}}\frac{g\left( t\right) - g\left( 0\right) }{t} = {f}^{\prime }\left( {\bar{y};y - \bar{y}}\right) . \] Therefore (4) holds, since \( y \in Y \) is arbitrary. For \( Y \subset X \) and \( x \in X \), we denote by \( \left\lbrack {x, Y}\right\rbrack \) the convex hull of \( \{ x\} \cup Y \) : \[ \left\lbrack {x, Y}\right\rbrack \mathrel{\text{:=}} \{ {tx} + \left( {1 - t}\right) y : t \in \left\lbrack {0,1}\right\rbrack, y \in Y\} . \] (5) In particular, if \( Y = \{ y\} \), then \( \left\lbrack {x, Y}\right\rbrack \) is simply the closed line segment connecting \( x \) and \( y \) . Note that below we admit the possibility that \( x \in Y \) . 2.3. Theorem. Suppose \( Y \subset {\mathbb{R}}^{n} \) is closed, bounded, and convex. Let \( x \in \) \( {\mathbb{R}}^{n} \), and let \( f \in \mathcal{F} \) be Gâteaux differentiable on a neighborhood of \( \left\lbrack {x, Y}\right\rbrack \) . Then there exists \( z \in \left\lbrack {x, Y}\right\rbrack \) such that \[ \mathop{\min }\limits_{{{y}^{\prime } \in Y}}f\left( {y}^{\prime }\right) - f\left( x\right) \leq \left\langle {{f}^{\prime }\left( z\right), y - x}\right\rangle \forall y \in Y. \] (6) Proof. Set \[ r \mathrel{\text{:=}} \mathop{\min }\limits_{{{y}^{\prime } \in Y}}f\left( {y}^{\prime }\right) - f\left( x\right) \] (7) and let \( U \) be an open neighborhood of \( \left\lbrack {x, Y}\right\rbrack \) on which \( f \) is Gâteaux differentiable. Define \( g : \left\lbrack {0,1}\right\rbrack \times U \rightarrow \mathbb{R} \) by \[ g\left( {t, y}\right) = f\left( {x + t\left( {y - x}\right) }\right) - {rt}. \] (8) Since \( g \) is lower semicontinuous and \( \left\lbrack {x, Y}\right\rbrack \) is compact, we have \( r \neq - \infty \) and there exists \( \left( {\bar{t},\bar{y}}\right) \in \left\lbrack {0,1}\right\rbrack \times Y \) at which \( g \) attains its minimum over \( \left\lbrack {0,1}\right\rbrack \times Y \) . We will show that (6) holds with \( z \) given by \[ z \mathrel{\text{:=}} x + \bar{t}\left( {\bar{y} - x}\right) \in \left\lbrack {x, Y}\right\rbrack \] (9) unless \( \bar{t} = 1 \), in which case we set \( z = x \) . The multidirectional inequality (6) will be deduced from necessary conditions for the attainment of a minimum. Let us first assume that \( \bar{t} \in \left( {0,1}\right) \) . Then the function \[ t \mapsto g\left( {t,\bar{y}}\right) \] attains its minimum over \( t \in \left\lbrack {0,1}\right\rbrack \) at \( t = \bar{t} \), and thus has a vanishing derivative there. We calculate this derivative using (8) and (9) to conclude \[ 0 = {\left. \frac{\partial }{\partial t}\right| }_{t = \bar{t}}g\left( {t,\bar{y}}\right) = \left\langle {{f}^{\prime }\left( z\right) ,\bar{y} - x}\right\rangle - r. \] (10) On the other hand, the function \[ y \mapsto g\left( {\bar{t}, y}\right) \] attains its minimum over \( y \in Y \) at \( \bar{y} \) . The necessary conditions in Proposition 2.2 state that \[ \left\langle {{\left. \frac{\partial }{\partial y}\right| }_{y = \bar{y}}g\left( {\bar{t}, y}\right), y - \bar{y}}\right\rangle \geq 0\forall y \in Y \] (11) It is easily calculated using (8) and (9) again that \[ {\left. \frac{\partial }{\partial y}\right| }_{y = \bar{y}}g\left( {\bar{t}, y}\right) = \bar{t}{f}^{\prime }\left( z\right) \] Substituting this value into (11) and dividing by \( \bar{t} > 0 \) gives us that \[ \left\langle {{f}^{\prime }\left( z\right), y - \bar{y}}\right\rangle \geq 0\forall y \in Y. \] (12) We now combine (10) and (12) to conclude that \[ r \leq \left\langle {{f}^{\prime }\left( z\right) ,\bar{y} - x}\right\rangle + \left\langle {{f}^{\prime }\left( z\right), y - \bar{y}}\right\rangle = \left\langle {{f}^{\prime }\left( z\right), y - x}\right\rangle \] which holds for all \( y \in Y \) . This statement is precisely (6) and finishes the proof of the theorem in this instance. The case \( \bar{t} = 0 \) is handled more simply than above. For each \( \left( {t, y}\right) \in \) \( (0,1\rbrack \times Y \), we note from (8) and the definition of a minimum that \[ f\left( x\right) = g\left( {0,\bar{y}}\right) \leq g\left( {t, y}\right) = f\left( {x + t\left( {y - x}\right) }\right) - {rt}. \] (13) After dividing (13) by \( t \) and rearranging terms, we have for all \( y \in Y \) that \[ r \leq \mathop{\lim }\limits_{{t \downarrow 0}}\frac{1}{t}\left( {f\left( {x + t\left( {y - x}\right) }\right) - f\left( x\right) }\right) = \left\langle {{f}^{\prime }\left( z\right), y - x}\right\rangle , \] since \( z = x \), and this is the same as (6). Finally we consider the case in which \( \bar{t} = 1 \) . We claim \( g\left( {1,\bar{y}}\right) = f\left( x\right) \) . Indeed, \( g\left( {1,\bar{y}}\right) \leq g\left( {0,\bar{y}}\right) = f\left( x\right) \), while on the other hand, \[ g\left( {1,\bar{y}}\right) = f\left( \bar{y}\right) - r = f\left( \bar{y}\right) - \mathop{\min }\limits_{{y \in Y}}f\left( y\right) + f\left( x\right) \geq f\left( x\right) . \] It now follows that \( f\left( x\right) \) is the minimum value of \( g \) on \( \left\lbrack {0,1}\right\rbrack \times Y \) . But \( f\left( x\right) = g\left( {0, y}\right) \) for any \( y \in Y \), and thus \( \left( {0, y}\right) \) is also a minimizer of \( g \) . Thus if \( \bar{t} = 1 \), then we equally could have chosen the minimizer with \( \bar{t} = 0 \), and we finish by deferring to that case, treated above. 2.4. Exercise. (a) Prove that the point \( z \) in Theorem 2.3 also satisfies \[ f\left( z\right) \leq \mathop{\min }\limits_{{w \in \left\lbrack {x, Y}\right\rbrack }}f\left( w\right) + \left| r\right| \] where \( r \) is given by (7). (b) Consider the special case of Theorem 2.3 in which \( n = 2, x = \) \( \left( {0,0}\right) \), and \[ Y = \{ \left( {1, t}\right) : - 1 \leq t \leq 1\} . \] Suppose that \( f\left( {0,0}\right) = 0 \) and that \( f\left( {1, t}\right) = 1\forall t \in \left\lbrack {-1,1}\right\rbrack \) . Prove that some point \( \left( {u, v}\right) \) in the tr
1012_(GTM178)Nonsmooth Analysis and Control Theory
41
\( f\left( x\right) \) is the minimum value of \( g \) on \( \left\lbrack {0,1}\right\rbrack \times Y \) . But \( f\left( x\right) = g\left( {0, y}\right) \) for any \( y \in Y \), and thus \( \left( {0, y}\right) \) is also a minimizer of \( g \) . Thus if \( \bar{t} = 1 \), then we equally could have chosen the minimizer with \( \bar{t} = 0 \), and we finish by deferring to that case, treated above. 2.4. Exercise. (a) Prove that the point \( z \) in Theorem 2.3 also satisfies \[ f\left( z\right) \leq \mathop{\min }\limits_{{w \in \left\lbrack {x, Y}\right\rbrack }}f\left( w\right) + \left| r\right| \] where \( r \) is given by (7). (b) Consider the special case of Theorem 2.3 in which \( n = 2, x = \) \( \left( {0,0}\right) \), and \[ Y = \{ \left( {1, t}\right) : - 1 \leq t \leq 1\} . \] Suppose that \( f\left( {0,0}\right) = 0 \) and that \( f\left( {1, t}\right) = 1\forall t \in \left\lbrack {-1,1}\right\rbrack \) . Prove that some point \( \left( {u, v}\right) \) in the triangle \( \left\lbrack {0, Y}\right\rbrack \) satisfies \[ {f}_{x}\left( {u, v}\right) \geq 1 + \left| {{f}_{y}\left( {u, v}\right) }\right| \] (c) With \( x \) and \( Y \) as in part (b), and with \( f\left( {x, y}\right) \mathrel{\text{:=}} x - y + {xy} \) , find all the points \( z \) that satisfy the conclusion of Theorem 2.3. ## The General Case of the Mean Value Inequality We now wish to extend the theorem to the case in which \( Y \) is a closed, bounded, convex subset of the Hilbert space \( X \), and \( f \) is merely lower semicontinuous. The first technical difficulty in proving the theorem in this greater generality is the same as that previously encountered in the proof of the Density Theorem. Namely, a lower semicontinuous function that is bounded below may not attain its minimum over a closed bounded set. We may note that the proof of Theorem 2.3 above is valid in infinite dimensions without modification if in addition \( Y \) is assumed to be compact, or \( f \) is assumed to be weakly lower semicontinuous. This is because these hypotheses imply the existence of the minimizer \( \left( {\bar{t},\bar{y}}\right) \) . But the additional assumptions are not actually required to prove the theorem in infinite dimensions, as we will see. The idea is to follow the above proof but to obtain the minimizer via a minimization principle (Theorem 1.4.1). However, there is a price to pay for this device to work effectively: The statement of the result itself needs a slight modification, a certain slackening of the inequality in the conclusion. The result is not true without this modification, and it is essentially required in the proof to avoid the possibility that \( \bar{t} = 1 \) . (We may check that if \( \bar{t} = 1 \), no relevant information is obtained from the necessary conditions.) In the above proof, the case \( \bar{t} = 1 \) was handled by deferring to the case \( \bar{t} = 0 \), but after adding the perturbation term, this is no longer possible. The second technical difficulty arises in replacing derivatives by proximal subgradients. When \( f \) is nondifferentiable, Proposition 2.2 is no longer available to us, and in the minimization process we have recourse instead to a smooth penalty term to replace the constraint that \( z \) lie in \( \left\lbrack {x, Y}\right\rbrack \) . This will now allow the minimizer \( z \) to fail to lie in \( \left\lbrack {x, Y}\right\rbrack \), although it can be brought arbitrarily near to that set (an example will show that this tolerance is necessary). Further, the quantity \( r \) defined by (7) must be replaced by one that is more stable under such approximations, namely \[ \widehat{r} \mathrel{\text{:=}} \mathop{\lim }\limits_{{\delta \downarrow 0}}\mathop{\inf }\limits_{{w \in Y + {\delta B}}}\{ f\left( w\right) - f\left( x\right) \} . \] (14) 2.5. Exercise. If \( X \) is finite dimensional, show that the \( \widehat{r} \) defined by (14) reduces to the \( r \) defined by (7). (Examples can be adduced to show that this may fail in infinite dimensions.) 2.6. Theorem (Mean Value Inequality). Let \( Y \) be a closed, convex, bounded subset of \( X \), and let \( x \in \operatorname{dom}f \), where \( f \in \mathcal{F} \) . Suppose that \( \widehat{r} \) defined above is not \( - \infty \), and let any \( \bar{r} < \widehat{r} \) and \( \varepsilon > 0 \) be given. Then there exist \( z \in \left\lbrack {x, Y}\right\rbrack + {\varepsilon B} \) and \( \zeta \in {\partial }_{P}f\left( z\right) \) such that \[ \bar{r} < \langle \zeta, y - x\rangle \forall y \in Y. \] Further, we can choose \( z \) to satisfy \[ f\left( z\right) < \mathop{\inf }\limits_{\left\lbrack x, Y\right\rbrack }f + \left| \bar{r}\right| + \varepsilon . \] We remark that the case \( \widehat{r} = \infty \) is not excluded; it can only arise if \( Y \cap \) \( \operatorname{dom}f = \varnothing \) . ## Proof of the Mean Value Inequality Proof. Without loss of generality, we assume that \( x = 0 \) and \( f\left( x\right) = 0 \) . Suppose \( \bar{r} < \widehat{r} \) and \( \varepsilon > 0 \) . We will first fix some constants that will be featured in the estimates to come. The hypotheses of the theorem justify the existence of such choices. We write \( \parallel Y\parallel \) for the quantity \( \sup \{ \parallel y\parallel : y \in Y\} \) . Let \( r \in \mathbb{R} \) and positive constants \( \delta, M, k \) be chosen so that \[ \bar{r} < r < \widehat{r},\;\left| r\right| < \left| \bar{r}\right| + \varepsilon \] (15) \[ \delta \leq \varepsilon \;\text{ and }\;y \in Y + {\delta B} \Rightarrow f\left( y\right) \geq r + \delta , \] (16) \[ {z}^{\prime } \in \left\lbrack {0, Y}\right\rbrack + \delta \bar{B} \Rightarrow f\left( {z}^{\prime }\right) \geq - M \] (17) \[ k > \frac{1}{{\delta }^{2}}\{ M + \left| r\right| + 1 + 2\parallel Y\parallel + \delta \} . \] (18) Now we define \( g : \left\lbrack {0,1}\right\rbrack \times X \times X \rightarrow ( - \infty ,\infty \rbrack \) by \[ g\left( {t, y, z}\right) \mathrel{\text{:=}} f\left( z\right) + k\parallel {ty} - z{\parallel }^{2} - {tr}. \] To simplify the notation somewhat, let us set \[ S \mathrel{\text{:=}} Y \times \{ \left\lbrack {0, Y}\right\rbrack + \delta \bar{B}\} \] and write \( s = \left( {y, z}\right) \) for elements in \( S \) . We can easily verify that \[ t \mapsto \mathop{\inf }\limits_{{s \in S}}g\left( {t, s}\right) \] (19) is continuous on \( \left\lbrack {0,1}\right\rbrack \), and thus attains its infimum at some point \( \bar{t} \in \left\lbrack {0,1}\right\rbrack \) . The choice of a large \( k \) value forces \( \bar{t} \) away from 1, which is the assertion of the first claim. Claim 1. \( \bar{t} < 1 \) . Proof. Let \( s = \left( {y, z}\right) \in S \), and note that if \( \parallel y - z\parallel \leq \delta \), then \( f\left( z\right) \geq r + \delta \) by (16). A lower bound for \( g\left( {1, s}\right) \) can be obtained as follows: \[ g\left( {1, s}\right) = f\left( z\right) + k\parallel y - z{\parallel }^{2} - r \] \[ \geq \left\{ \begin{array}{ll} f\left( z\right) + k\parallel y - z{\parallel }^{2} - r & \text{ if }\parallel y - z\parallel \leq \delta \\ f\left( z\right) + k{\delta }^{2} - r & \text{ if }\parallel y - z\parallel > \delta \end{array}\right. \] \[ \geq \min \{ r + \delta - r, - M + \left( {M + \left| r\right| + 1}\right) - r\} > 0. \] On the other hand, we have an upper bound of the function in (19) by setting \( t = 0 \) and noting that \[ \mathop{\inf }\limits_{{s \in S}}g\left( {0, s}\right) \leq g\left( {0, y,0}\right) = f\left( 0\right) + k\parallel 0{\parallel }^{2} = 0. \] (20) Thus the infimum of \( t \mapsto \mathop{\inf }\limits_{{s \in S}}g\left( {t, s}\right) \) taken over \( t \in \left\lbrack {0,1}\right\rbrack \) cannot occur at \( t = 1 \), and we have \( \bar{t} < 1 \) . We now address the difficulty that the minimum of \( s \mapsto g\left( {\bar{t}, s}\right) \) taken over \( s \in S \) may not be attained at any point in \( S \) . We resort to a minimization principle to obtain a minimum of \( g\left( {\bar{t}, \cdot }\right) \) perturbed by a linear function. Note that \( g\left( {\bar{t}, \cdot }\right) \in \mathcal{F} \) and is bounded below on the closed bounded set \( S \) . By Theorem 1.4.1, for each small \( \eta > 0 \), there exists \( \xi \mathrel{\text{:=}} \left( {{\xi }_{y},{\xi }_{z}}\right) \in X \times X \) and \( \bar{s} \mathrel{\text{:=}} \left( {\bar{y},\bar{z}}\right) \in S \) so that \( \parallel \xi \parallel < \eta \) and the function \[ s \mapsto g\left( {\bar{t}, s}\right) + \langle \xi, s\rangle \] (21) attains its minimum over \( s \in S \) at \( \bar{s} \) . Of course \( \xi \) and \( \bar{s} \) depend on \( \eta \), as does the choice of \( \zeta \) appearing in the conclusions of the theorem, the latter being defined as \[ \zeta \mathrel{\text{:=}} {2k}\left( {\bar{t}\bar{y} - \bar{z}}\right) - {\xi }_{z} \] We will see that the conclusions of the theorem hold for these choices of \( \zeta \) and \( \bar{z} \), provided that \( \eta \) is chosen sufficiently small. Claim 2. If \( \eta \) is small enough, we have \( \bar{z} \in \left\lbrack {0, Y}\right\rbrack + {\delta B} \) . Proof. Suppose on the contrary that \( {d}_{\left\lbrack 0, Y\right\rbrack }\left( \bar{z}\right) \geq \delta \) . Observe first that \[ \left| {\langle \xi, s\rangle }\right| \leq \eta \left( {2\parallel Y\parallel + \delta }\right) \forall s \in S. \] (22) Then since \( \bar{t}\bar{y} \in \left\lbrack {0, Y}\right\rbrack \), we have by (22) and (17) that \[ g\left( {\bar{t},\bar{s}}\right) + \langle \xi ,\bar{s}\rangle = f\left( \bar{z}\right) + k\parallel \bar{t}\bar{y} - \bar{z}{\parallel }^{2} - r\bar{t} + \langle \xi ,\bar{s}\rangle \] \[ \geq - M + k{\delta }^{2} - \left| r\right| - \eta \left( {2\parallel Y\parallel + \delta }\right) , \] which by the choice of \( k \) in (18) can be made greater than 1 if \( \eta \) is sufficiently s
1012_(GTM178)Nonsmooth Analysis and Control Theory
42
f the theorem hold for these choices of \( \zeta \) and \( \bar{z} \), provided that \( \eta \) is chosen sufficiently small. Claim 2. If \( \eta \) is small enough, we have \( \bar{z} \in \left\lbrack {0, Y}\right\rbrack + {\delta B} \) . Proof. Suppose on the contrary that \( {d}_{\left\lbrack 0, Y\right\rbrack }\left( \bar{z}\right) \geq \delta \) . Observe first that \[ \left| {\langle \xi, s\rangle }\right| \leq \eta \left( {2\parallel Y\parallel + \delta }\right) \forall s \in S. \] (22) Then since \( \bar{t}\bar{y} \in \left\lbrack {0, Y}\right\rbrack \), we have by (22) and (17) that \[ g\left( {\bar{t},\bar{s}}\right) + \langle \xi ,\bar{s}\rangle = f\left( \bar{z}\right) + k\parallel \bar{t}\bar{y} - \bar{z}{\parallel }^{2} - r\bar{t} + \langle \xi ,\bar{s}\rangle \] \[ \geq - M + k{\delta }^{2} - \left| r\right| - \eta \left( {2\parallel Y\parallel + \delta }\right) , \] which by the choice of \( k \) in (18) can be made greater than 1 if \( \eta \) is sufficiently small. On the other hand, recall that \( \bar{s} \) achieves the infimum in the function (21). Therefore if \( \eta \leq {\left( 2\parallel Y\parallel + \delta \right) }^{-1}/2 \), we have by (20) and (22) \[ g\left( {\bar{t},\bar{s}}\right) + \langle \xi ,\bar{s}\rangle \leq g\left( {0,\bar{y},0}\right) + \langle \xi ,\bar{s}\rangle \] \[ \leq \eta \left( {2\parallel \bar{y}\parallel + \delta }\right) \] \[ \leq 1/2\text{.} \] Thus we conclude \( \bar{z} \in \left\lbrack {0, Y}\right\rbrack + {\delta B} \) as claimed. Claim 3. \( \zeta \in {\partial }_{P}f\left( \bar{z}\right) \) . Proof. Since \( z \mapsto f\left( z\right) + k\parallel \bar{t}\bar{y} - z{\parallel }^{2} + \left\langle {{\xi }_{z}, z}\right\rangle \) has a minimum over \( \left\lbrack {0, Y}\right\rbrack + \delta \bar{B} \) at \( z = \bar{z} \), and \( \bar{z} \) is a point in the interior of this set (Claim 2), we have that \[ \zeta = - {\left( k\parallel \bar{t}\bar{y} - \left( \cdot \right) {\parallel }^{2} + \left\langle {\xi }_{z},\left( \cdot \right) \right\rangle \right) }^{\prime }\left( \bar{z}\right) \in {\partial }_{P}f\left( \bar{z}\right) . \] We next turn to proving the upper bound on \( f\left( \bar{z}\right) \) . Claim 4. The estimate \( f\left( \bar{z}\right) \leq \mathop{\inf }\limits_{{z \in \left\lbrack {0, Y}\right\rbrack }}f\left( z\right) + \left| r\right| + \varepsilon \) holds, provided \( \eta \) is small enough. Proof. Using (22), we calculate \[ f\left( \bar{z}\right) - \bar{t}r \leq f\left( \bar{z}\right) + k\parallel \bar{t}\bar{y} - \bar{z}{\parallel }^{2} - \bar{t}r + \langle \xi ,\left( {\bar{y},\bar{z}}\right) \rangle + \eta \left( {2\parallel Y\parallel + \delta }\right) \] \[ = \mathop{\inf }\limits_{{s \in S}}\{ g\left( {\bar{t}, s}\right) + \langle \xi, s\rangle \} + \eta \left( {2\parallel Y\parallel + \delta }\right) \] \[ \leq \mathop{\inf }\limits_{{\left( {y, z}\right) \in S}}\{ f\left( z\right) + k\parallel {ty} - z{\parallel }^{2} - {tr} + \left\langle {\xi ,\left( {y, z}\right) }\right\rangle \} + \eta \left( {2\parallel Y\parallel + \delta }\right) \] \[ \leq \mathop{\inf }\limits_{{z \in \left\lbrack {0, Y}\right\rbrack }}f\left( z\right) + \max \{ 0, - r\} + {2\eta }\left( {2\parallel Y\parallel + \delta }\right) . \] The final inequality is justified since the last infimum is only over \( z \in \left\lbrack {0, Y}\right\rbrack \) , and included in the next to last infimum is the case when \( {ty} = z \in \left\lbrack {0, Y}\right\rbrack \) , making it therefore an infimum over a larger set. Hence \[ f\left( \bar{z}\right) \leq \mathop{\inf }\limits_{{z \in \left\lbrack {0, Y}\right\rbrack }}f\left( z\right) + \max \{ \bar{t},1 - \bar{t}\} \left| r\right| + {2\eta }\left( {2\parallel Y\parallel + \delta }\right) \] \[ \leq \mathop{\inf }\limits_{{z \in \left\lbrack {0, Y}\right\rbrack }}f\left( z\right) + \left| \bar{r}\right| + \varepsilon \] provided \( \eta \) is sufficiently small, where the second relation in (15) has been used. Claim 5. Suppose \( \bar{t} = 0 \), and \( 0 < \eta < {\left\lbrack \left( r - \bar{r}\right) /6\left( 2\parallel Y\parallel + \delta \right) \right\rbrack }^{2} \) . Then \( \bar{r} < \langle \zeta, y\rangle \) for all \( y \in Y \) . Proof. In the present case with \( \bar{t} = 0 \), the definition of \( \zeta \) reduces to \( \zeta = \) \( - {2k}\bar{z} - {\xi }_{z} \) . Let us note (using (22)) that \[ \mathop{\inf }\limits_{{s \in S}}g\left( {0, s}\right) \geq \mathop{\inf }\limits_{{s \in S}}\{ g\left( {0, s}\right) + \langle \xi, s\rangle \} - \eta \left( {2\parallel Y\parallel + \delta }\right) \] \[ = f\left( \bar{z}\right) + k\parallel \bar{z}{\parallel }^{2} + \left\langle {\xi ,\left( {\bar{y},\bar{z}}\right) }\right\rangle - \eta \left( {2\parallel Y\parallel + \delta }\right) \] (23) \[ \geq f\left( \bar{z}\right) + k\parallel \bar{z}{\parallel }^{2} - {2\eta }\left( {2\parallel Y\parallel + \delta }\right) . \] Now we let \( t \mathrel{\text{:=}} \min \left\{ {\sqrt{\eta },\left( {r - \bar{r}}\right) /{3k}\left( {\parallel Y{\parallel }^{2} + \delta }\right) }\right\} > 0 \) . We have for any \( y \in Y \) that \[ \mathop{\inf }\limits_{{s \in S}}g\left( {t, s}\right) \leq f\left( \bar{z}\right) + k\parallel {ty} - \bar{z}{\parallel }^{2} - {tr} \] (24) Since \( \bar{t} = 0 \) by definition provides a minimum of (19), it follows from (23) and (24) that for all \( y \in Y \) \[ 0 \leq \mathop{\inf }\limits_{{s \in S}}g\left( {t, s}\right) - \mathop{\inf }\limits_{{s \in S}}g\left( {0, s}\right) \] \[ \leq k\parallel {ty} - \bar{z}{\parallel }^{2} - {tr} - k\parallel \bar{z}{\parallel }^{2} + {2\eta }\left( {2\parallel Y\parallel + \delta }\right) \] \[ = t\left\lbrack {\left\langle {-{2k}\bar{z}, y}\right\rangle - r}\right\rbrack + {t}^{2}k\parallel y{\parallel }^{2} + {2\eta }\left( {2\parallel Y\parallel + \delta }\right) \] \[ = t\left\lbrack {\langle \zeta, y\rangle - r}\right\rbrack + {t}^{2}k\parallel y{\parallel }^{2} + t\langle {\xi }_{z}, y\rangle + {2\eta }\left( {2\parallel Y\parallel + \delta }\right) \] \[ \leq t\left\lbrack {\langle \zeta, y\rangle - r}\right\rbrack + {t}^{2}k\parallel Y{\parallel }^{2} + {t\eta }\parallel Y\parallel + {2\eta }\left( {2\parallel Y\parallel + \delta }\right) . \] Dividing through by \( t \) and using our choices of \( t \) and \( \eta \), we obtain \[ \langle \zeta, y\rangle \geq r - {tk}\parallel Y{\parallel }^{2} - \eta \parallel Y\parallel - \frac{2\eta }{t}\left( {2\parallel Y\parallel + \delta }\right) \] \[ > \bar{r} \] which concludes the proof of this claim. To handle the situation where \( \bar{t} > 0 \), we first prove an estimate regarding \( \bar{y} \) . Claim 6. If \( \eta \) is sufficiently small, then \( \langle \zeta ,\bar{y}\rangle > \left( {r + \bar{r}}\right) /2 \) . Proof. In a manner similar to the beginning of the proof of Claim 5, we have \[ \mathop{\inf }\limits_{{s \in S}}g\left( {\bar{t}, s}\right) \geq f\left( \bar{z}\right) + k\parallel \bar{t}\bar{y} - \bar{z}{\parallel }^{2} - {2\eta }\left( {2\parallel Y\parallel + \delta }\right) , \] (25) and \[ \mathop{\inf }\limits_{{s \in S}}g\left( {1, s}\right) \leq f\left( \bar{z}\right) + k\parallel \left( {1 - \bar{t}}\right) \bar{y} + \left( {\bar{t}\bar{y} - \bar{z}}\right) {\parallel }^{2} - r. \] (26) Recall that \( \bar{t} \) provides an infimum of the function in (21). By subtracting (25) from (26), we obtain \[ 0 \leq \mathop{\inf }\limits_{{s \in S}}g\left( {1, s}\right) - \mathop{\inf }\limits_{{s \in S}}g\left( {\bar{t}, s}\right) \] \[ \leq k{\begin{Vmatrix}\left( 1 - \bar{t}\right) \bar{y} + \left( \bar{t}\bar{y} - \bar{z}\right) \end{Vmatrix}}^{2} - r - k\parallel \bar{t}\bar{y} - \bar{z}{\parallel }^{2} + {2\eta }\left( {2\parallel Y\parallel + \delta }\right) \] \[ = \left( {1 - \bar{t}}\right) \left\lbrack {\left\langle {{2k}\left( {\bar{t}\bar{y} - \bar{z}}\right) ,\bar{y}}\right\rangle - r}\right\rbrack + {2\eta }\left( {2\parallel Y\parallel + \delta }\right) . \] This in turn implies \[ \langle {2k}\left( {\bar{t}\bar{y} - \bar{z}}\right) ,\bar{y}\rangle \geq r - \frac{2\eta }{1 - \bar{t}}\left( {2\parallel Y\parallel + \delta }\right) . \] Consequently we have \[ \langle \zeta ,\bar{y}\rangle = \left\langle {{2k}\left( {\bar{t}\bar{y} - \bar{z}}\right) - {\xi }_{z},\bar{y}}\right\rangle \] \[ \geq \left\langle {{2k}\left( {\bar{t}\bar{y} - \bar{z}}\right) ,\bar{y}}\right\rangle - \eta \parallel Y\parallel \] \[ \geq r - \eta \left\lbrack {\frac{2}{1 - \bar{t}}\left( {2\parallel Y\parallel + \delta }\right) + \parallel Y\parallel }\right\rbrack \] Hence \( \langle \zeta ,\bar{y}\rangle > \left( {r + \bar{r}}\right) /2 \) if \( \eta \) is chosen small enough. The proof of the theorem will ensue from the next claim. Claim 7. If in addition to the requirements on \( \eta \) in all of the earlier claims, we also have \( \eta < \bar{t}\left( {r - \bar{r}}\right) /\left( {8\parallel Y\parallel }\right) \), then \[ \bar{r} < \langle \zeta, y\rangle \forall y \in Y. \] Proof. We have that the function \[ y \mapsto k\parallel \bar{t}y - \bar{z}{\parallel }^{2} + \left\langle {{\xi }_{y}, y}\right\rangle \] achieves its infimum over \( y \in Y \) at \( y = \bar{y} \) . Since \( Y \) is convex, we have by Proposition 2.2 that for all \( y \in Y \) , \[ \left\langle {\frac{{\zeta }_{y}}{\bar{t}}, y - \bar{y}}\right\rangle \geq - 2\frac{\eta }{\bar{t}}\parallel Y\parallel \] Therefore \[ \langle \zeta, y - \bar{y}\rangle = \left\langle {{2k}\left( {\bar{t}\bar{y} - \bar{z}}\right) - {\xi }_{z}, y - \bar{y}}\right\rangle \] \[ \geq - 2\frac{\eta }{\bar{t}}\parallel Y\parallel - {2\eta }\parallel Y\parallel \] \[ \geq - 4\frac{\eta }{\bar{t}}\parallel Y\parallel \geq \frac{\bar{r} - r}{2} \] Finally, from this and Claim 6 we conclude that for any \( y \in Y \) that \[ \langle \zeta, y\rangle \geq \langle \zeta ,\bar{y}\rangle - \frac{r - \bar{r}}{2} \] \[ > \frac{r + \bar{r}
1012_(GTM178)Nonsmooth Analysis and Control Theory
43
ght) \), then \[ \bar{r} < \langle \zeta, y\rangle \forall y \in Y. \] Proof. We have that the function \[ y \mapsto k\parallel \bar{t}y - \bar{z}{\parallel }^{2} + \left\langle {{\xi }_{y}, y}\right\rangle \] achieves its infimum over \( y \in Y \) at \( y = \bar{y} \) . Since \( Y \) is convex, we have by Proposition 2.2 that for all \( y \in Y \) , \[ \left\langle {\frac{{\zeta }_{y}}{\bar{t}}, y - \bar{y}}\right\rangle \geq - 2\frac{\eta }{\bar{t}}\parallel Y\parallel \] Therefore \[ \langle \zeta, y - \bar{y}\rangle = \left\langle {{2k}\left( {\bar{t}\bar{y} - \bar{z}}\right) - {\xi }_{z}, y - \bar{y}}\right\rangle \] \[ \geq - 2\frac{\eta }{\bar{t}}\parallel Y\parallel - {2\eta }\parallel Y\parallel \] \[ \geq - 4\frac{\eta }{\bar{t}}\parallel Y\parallel \geq \frac{\bar{r} - r}{2} \] Finally, from this and Claim 6 we conclude that for any \( y \in Y \) that \[ \langle \zeta, y\rangle \geq \langle \zeta ,\bar{y}\rangle - \frac{r - \bar{r}}{2} \] \[ > \frac{r + \bar{r}}{2} - \frac{r - \bar{r}}{2} \] \[ = \bar{r} \] which completes the proof of the theorem. 2.7. Exercise. (a) Show that if \( x \in Y \), then the point \( z \) in Theorem 2.6 satisfies \( f\left( z\right) < f\left( x\right) + \varepsilon \) . (b) Deduce from Theorem 2.6 the (unidirectional) Proximal Mean Value Theorem: If \( f \in \mathcal{F} \) and \( x, y \in \operatorname{dom}f \) are given, then for any \( \varepsilon > 0 \) there exists \( z \in \left\lbrack {x, y}\right\rbrack + {\varepsilon B} \) such that, for some \( \zeta \in {\partial }_{P}f\left( z\right) \) we have \[ f\left( y\right) - f\left( x\right) < \langle \zeta, y - x\rangle + \varepsilon . \] (c) If now the point \( y \) in part (b) is not in \( \operatorname{dom}f \), deduce that there are points \( z \) arbitrarily near \( \left\lbrack {x, y}\right\rbrack \) admitting proximal subgradients \( \zeta \in {\partial }_{P}f\left( z\right) \) for which \( \langle \zeta, y - x\rangle \) (and hence \( \parallel \zeta \parallel \) ) is arbitrarily large. (d) Let \( f \) be Lipschitz on a neighborhood of \( \left\lbrack {x, y}\right\rbrack \) . Prove the existence of \( z \in \left\lbrack {x, y}\right\rbrack \) and \( \zeta \in {\partial }_{L}f\left( z\right) \) such that \[ f\left( y\right) - f\left( x\right) \leq \langle \zeta, y - x\rangle . \] ## The Decrease Principle As the first of several applications of the Mean Value Inequality we derive a result which plays an important role in the next section. To motivate it, consider a differentiable function \( f \) whose derivative at \( {x}_{0} \) is nonvanishing: \( {f}^{\prime }\left( {x}_{0}\right) \neq 0 \) . Since \( {x}_{0} \) cannot therefore be a local minimum of \( f \), we conclude that for any open ball \( {x}_{0} + {\rho B} \) around \( {x}_{0} \), we must have \[ \mathop{\inf }\limits_{{x \in {x}_{0} + {\rho B}}}f\left( x\right) < f\left( {x}_{0}\right) . \] This fact can be extended locally and quantified, as follows: 2.8. Theorem. Let \( f \in \mathcal{F} \) . Suppose that for some \( \delta > 0 \) and \( \rho > 0 \) we have the following condition: \[ z \in {x}_{0} + {\rho B},\;\zeta \in {\partial }_{P}f\left( z\right) \Rightarrow \parallel \zeta \parallel \geq \delta . \] Then \[ \mathop{\inf }\limits_{{x \in {x}_{0} + {\rho B}}}f\left( x\right) \leq f\left( {x}_{0}\right) - {\rho \delta }. \] Proof. We may suppose that \( {x}_{0} \in \operatorname{dom}f \) . We will show that for any \( s \in \) \( \left( {0,\rho }\right) \), for all \( \varepsilon > 0 \) arbitrarily small, we have \[ \mathop{\inf }\limits_{{{x}_{0} + {\rho B}}}f \leq f\left( {x}_{0}\right) - {s\delta } + \varepsilon \] (27) a fact which implies the estimate of the theorem. Clearly there is no loss of generality in assuming that the left side of (27) is finite, since otherwise the conclusion of the theorem is immediate. Given \( s \in \left( {0,\rho }\right) \), set \( Y \mathrel{\text{:=}} {x}_{0} + s\bar{B} \) , and observe that the corresponding quantity \( \widehat{r} \) defined via (14) satisfies (for \( x \mathrel{\text{:=}} {x}_{0}) \) \[ \infty > \widehat{r} \geq \mathop{\inf }\limits_{{{x}_{0} + {\rho B}}}f - f\left( {x}_{0}\right) \] Choose \( \varepsilon \in \left( {0,\rho - s}\right) \), and set \( \bar{r} \mathrel{\text{:=}} \widehat{r} - \varepsilon \) . Applying Theorem 2.6 gives \[ \mathop{\inf }\limits_{{{x}_{0} + {\rho B}}}f - f\left( {x}_{0}\right) \leq \mathop{\inf }\limits_{{{x}_{0} + {sB}}}f - f\left( {x}_{0}\right) \] \[ \leq \widehat{r} = \bar{r} + \varepsilon \] \[ < \langle \zeta ,{su}\rangle + \varepsilon \forall u \in \bar{B} \] where \( \zeta \in {\partial }_{P}f\left( z\right) \) and \( z \in Y + {\varepsilon B} = {x}_{0} + s\bar{B} + {\varepsilon B} \subset {x}_{0} + {\rho B} \) . By hypothesis, we have \( \parallel \zeta \parallel \geq \delta > 0 \) . Choosing \( u = - \zeta /\parallel \zeta \parallel \) in the preceding inequality therefore gives precisely (27). Theorem 2.8 implies a hybrid form of Theorems 1.4.1 and 1.4.2, the two minimization principles proved in Chapter 1. 2.9. Corollary. Let \( f \in \mathcal{F} \) be bounded below, and let \( x \in X,\varepsilon > 0 \) be such that \[ f\left( x\right) < \mathop{\inf }\limits_{X}f + \varepsilon . \] Then for any \( \lambda > 0 \) there exists \( z \in x + {\lambda B} \) with \( f\left( z\right) < \mathop{\inf }\limits_{X}f + \varepsilon \) and \( \zeta \in {\partial }_{P}f\left( z\right) \) such that \( \parallel \zeta \parallel < \varepsilon /\lambda \) . 2.10. Exercise. Prove the corollary. Deduce also that for some \( \sigma \geq \) 0 , the function \[ {x}^{\prime } \mapsto f\left( {x}^{\prime }\right) - \left\langle {\zeta ,{x}^{\prime }}\right\rangle + \sigma {\begin{Vmatrix}{x}^{\prime } - z\end{Vmatrix}}^{2} \] attains a local minimum at \( {x}^{\prime } = z \) . Compare this conclusion to those of Theorems 1.4.1 and 1.4.2. Note also that if \( f \) is differentiable then we get \( \begin{Vmatrix}{{f}^{\prime }\left( z\right) }\end{Vmatrix} \leq \sqrt{\varepsilon } \) by taking \( \lambda = \sqrt{\varepsilon } \), which improves upon Exercise 1.4.3. ## Strong Versus Weak Monotonicity Let \( C \) be a nonempty compact convex subset of \( X \) . A function \( f \) is said to be strongly decreasing relative to \( C \) if \[ \forall t > 0,\forall y \in x + {tC}\text{, we have}f\left( y\right) \leq f\left( x\right) \text{.} \] We say that \( f \) is weakly decreasing relative to \( C \) if \[ \forall t > 0\;\exists y \in x + {tC}\text{ such that }f\left( y\right) \leq f\left( x\right) . \] The lower and upper support functions of \( C,{h}_{C}\left( \cdot \right) \), and \( {H}_{C}\left( \cdot \right) \), respectively, are defined as follows: \[ {h}_{C}\left( \zeta \right) \mathrel{\text{:=}} \min \{ \langle \zeta, c\rangle : c \in C\} ,\;{H}_{C}\left( \zeta \right) \mathrel{\text{:=}} \max \{ \langle \zeta, c\rangle : c \in C\} . \] The following characterizations of weak and strong decrease foreshadow certain corresponding properties of control trajectories that figure prominently in Chapter 4. 2.11. Theorem. Let \( f \in \mathcal{F} \) . Then \( f \) is strongly decreasing relative to \( C \) \( {iff} \) \[ {H}_{C}\left( \zeta \right) \leq 0\forall \zeta \in {\partial }_{P}f\left( x\right) ,\forall x \in X \] \( f \) is weakly decreasing relative to \( C \) iff \[ {h}_{C}\left( \zeta \right) \leq 0\forall \zeta \in {\partial }_{P}f\left( x\right) ,\forall x \in X. \] Proof. The condition involving \( {H}_{C} \) can be expressed more compactly in the form \[ {H}_{C}\left( {{\partial }_{P}f\left( x\right) }\right) \leq 0\forall x. \] Suppose that this holds, and let \( y \in x + {tC} \) for \( t > 0 \) . We wish to prove that \( f\left( y\right) \leq f\left( x\right) \) ; there is no loss of generality in assuming \( x \in \operatorname{dom}f \) . If \( y \in \) dom \( f \) too, then by the Proximal Mean Value Theorem (Exercise 2.7(b)), for any \( \varepsilon > 0 \) we can estimate \( f\left( y\right) - f\left( x\right) \) from above by a term of the form \( \langle \zeta, y - x\rangle + \varepsilon ,\zeta \in {\partial }_{P}f\left( z\right) \) . But \( y - x \in {tC} \), so that \( \langle \zeta, y - x\rangle \leq {H}_{C}\left( \zeta \right) \leq 0 \) by assumption. It follows that \( f\left( y\right) \leq f\left( x\right) \) . There remains the possibility that \( f\left( y\right) = + \infty \) . But then \( \langle \zeta, y - x\rangle \) can be made arbitrarily large by Exercise 2.7(c), which contradicts \( {H}_{C}\left( \zeta \right) \leq 0 \) . For the converse, let \( f \) be strongly decreasing and let \( \zeta \in {\partial }_{P}f\left( x\right) \) . For any \( c \in C \) and \( t > 0 \) we then have \( f\left( {x + {tc}}\right) \leq f\left( x\right) \) . Combined with the proximal subgradient inequality for \( \zeta \), this implies \( \langle \zeta, c\rangle \leq 0 \) . Thus \( {H}_{C}\left( {{\partial }_{P}f\left( x\right) }\right) \leq 0 \) . We now suppose that \( {h}_{C}\left( {{\partial }_{P}f\left( x\right) }\right) \leq 0\forall x \) . Let \( t > 0 \) be given, and apply Theorem 2.6 with \( Y \mathrel{\text{:=}} x + {tC} \) (compact) and \( \varepsilon > 0 \) . We obtain \[ \mathop{\min }\limits_{{x + {tC}}}f - f\left( x\right) \leq \langle \zeta ,{tc}\rangle + \varepsilon \forall c \in C. \] Now take the minimum over \( c \in C \) on the right side of this inequality. The result is \[ \mathop{\min }\limits_{{x + {tC}}}f - f\left( x\right) \leq t{h}_{C}\left( \zeta \right) + \varepsilon \] Since \( {h}_{C}\left( \zeta \right) \leq 0 \) by assumption, and since \( \varepsilon \) is arbitrary, we deduce the existence of \( y \in x + {tC} \) such that \( f\left( y\right) \leq f\left( x\right) \) ; i.e., \( f \) is weakly decreasing. Now suppose that \( f \) is weakly decreasing, and let \( \zeta \in {\partial }_{P}f\left( x\right) \) . For each \( t > 0 \) there exists \
1012_(GTM178)Nonsmooth Analysis and Control Theory
44
al }_{P}f\left( x\right) }\right) \leq 0 \) . We now suppose that \( {h}_{C}\left( {{\partial }_{P}f\left( x\right) }\right) \leq 0\forall x \) . Let \( t > 0 \) be given, and apply Theorem 2.6 with \( Y \mathrel{\text{:=}} x + {tC} \) (compact) and \( \varepsilon > 0 \) . We obtain \[ \mathop{\min }\limits_{{x + {tC}}}f - f\left( x\right) \leq \langle \zeta ,{tc}\rangle + \varepsilon \forall c \in C. \] Now take the minimum over \( c \in C \) on the right side of this inequality. The result is \[ \mathop{\min }\limits_{{x + {tC}}}f - f\left( x\right) \leq t{h}_{C}\left( \zeta \right) + \varepsilon \] Since \( {h}_{C}\left( \zeta \right) \leq 0 \) by assumption, and since \( \varepsilon \) is arbitrary, we deduce the existence of \( y \in x + {tC} \) such that \( f\left( y\right) \leq f\left( x\right) \) ; i.e., \( f \) is weakly decreasing. Now suppose that \( f \) is weakly decreasing, and let \( \zeta \in {\partial }_{P}f\left( x\right) \) . For each \( t > 0 \) there exists \( c \in C \) for which \( f\left( {x + {tc}}\right) \leq f\left( x\right) \) . Let \( {t}_{i} \downarrow 0 \) and extract a subsequence of the corresponding points \( {c}_{i} \) (without relabeling) so that they converge to \( c \in C \) . For all \( i \) sufficiently large, the proximal subgradient inequality gives \[ 0 \geq f\left( {x + {t}_{i}{c}_{i}}\right) - f\left( x\right) \geq \left\langle {\zeta ,{t}_{i}{c}_{i}}\right\rangle - \sigma {\begin{Vmatrix}{t}_{i}{c}_{i}\end{Vmatrix}}^{2}. \] Dividing by \( {t}_{i} \) and passing to the limit yields \( \langle \zeta, c\rangle \leq 0 \), whence \( {h}_{C}\left( \zeta \right) \leq \) 0. ## 2.12. Exercise. (a) Show that the characterization of weak decrease continues to hold for a set \( C \) which is closed, bounded, and convex (not necessarily compact) if the function \( f \) is assumed to be weakly lower semicontinuous. (b) Let \( X = {\mathbb{R}}^{2} \), and let \( C = \{ \left( {t,1 - t}\right) : 0 \leq t \leq 1\} \) . Under what condition on \( {f}^{\prime }\left( x\right) \) is a \( {C}^{1} \) function \( f : {\mathbb{R}}^{2} \rightarrow \mathbb{R} \) weakly decreasing relative to \( C \) ? ## 3 Solving Equations Consider an equation \[ f\left( {x,\alpha }\right) = 0 \] (1) where \( f \) is a given function of two variables. It is a familiar consideration in mathematics to seek to solve this equation for \( x \), while viewing \( \alpha \) as a parameter. Typically this is done in a neighborhood of a given point \( \left( {{x}_{0},{\alpha }_{0}}\right) \) for which (1) is satisfied, and the important issues are these: For a given \( \alpha \) sufficiently near to \( {\alpha }_{0} \), does there continue to be at least one value of \( x \) for which (1) holds? How does this set of solutions vary with \( \alpha \) ? When can we be sure there is a suitably "nice" function \( x\left( \alpha \right) \) which "tracks" solutions as \( \alpha \) varies; i.e., such that \( f\left( {x\left( \alpha \right) ,\alpha }\right) = 0 \), and \( x\left( {\alpha }_{0}\right) = {x}_{0} \) ? We will develop a simple and powerful way to deal with such issues via nonsmooth analysis, and in so doing recover some classical results as well as establish new ones. For a given nonnegative function \( f \), we are interested, for given \( \alpha \), in the points \( x \in X \) (a Hilbert space) for which (1) holds, and we wish to limit attention to just those \( x \) lying in a prescribed subset \( \Omega \) of \( X \) . We define, then, the feasible set \( \Phi \left( \alpha \right) \) as follows: \[ \Phi \left( \alpha \right) \mathrel{\text{:=}} \{ x \in \Omega : f\left( {x,\alpha }\right) = 0\} . \] (2) The following result is the key to our approach; in its statement, \( {\partial }_{P}f\left( {x,\alpha }\right) \) refers to the proximal subdifferential of the function \( x \mapsto f\left( {x,\alpha }\right) \) . 3.1. Theorem (Solvability Theorem). Let \( A \) be a parameter set, and suppose that for each \( \alpha \in A \), the function \( x \mapsto f\left( {x,\alpha }\right) \) is nonnegative and belongs to \( \mathcal{F}\left( X\right) \) . Suppose that for some \( \delta > 0 \) and subset \( V \) of \( X \), we have the following nonstationarity condition: \[ \alpha \in A,\;x \in V,\;f\left( {x,\alpha }\right) > 0,\;\zeta \in {\partial }_{P}f\left( {x,\alpha }\right) \Rightarrow \parallel \zeta \parallel \geq \delta . \] Then for all \( \left( {x,\alpha }\right) \in X \times A \) we have \[ \min \{ d\left( {x;\operatorname{comp}V}\right), d\left( {x;\operatorname{comp}\Omega }\right), d\left( {x;\Phi \left( \alpha \right) }\right) \} \leq \frac{f\left( {x,\alpha }\right) }{\delta }. \] (3) Proof. If the conclusion is false, then for some \( \left( {x,\alpha }\right) \in X \times A, f\left( {x,\alpha }\right) \) is finite, and there exists \( \rho > 0 \) such that \[ \min \{ d\left( {x;\operatorname{comp}V}\right), d\left( {x;\operatorname{comp}\Omega }\right), d\left( {x;\Phi \left( \alpha \right) }\right) \} > \rho > \frac{f\left( {x,\alpha }\right) }{\delta }. \] It follows that \[ x + {\rho B} \subset V,\;x + {\rho B} \subset \Omega ,\;d\left( {x;\Phi \left( \alpha \right) }\right) > \rho . \] Thus we have \( f\left( {{x}^{\prime },\alpha }\right) > 0 \) for all \( {x}^{\prime } \in x + {\rho B} \) . By hypothesis, then, the proximal subgradients of \( f\left( {\cdot ,\alpha }\right) \) on the open set \( x + {\rho B} \) have norm bounded below by \( \delta \) . We invoke the Decrease Principle (Theorem 2.8) to conclude \[ 0 \leq \mathop{\inf }\limits_{{{x}^{\prime } \in x + {\rho B}}}f\left( {{x}^{\prime },\alpha }\right) \leq f\left( {x,\alpha }\right) - {\rho \delta } < 0. \] This contradiction proves the theorem. We remark that the context in which the theorem is most often used is that in which a given point \( \left( {{x}_{0},{\alpha }_{0}}\right) \) has been identified at which \( f = 0 \), and in which \( V \) and \( \Omega \) are neighborhoods of \( {x}_{0} \) . Then (3) asserts that \[ d\left( {x;\Phi \left( \alpha \right) }\right) \leq \frac{f\left( {x,\alpha }\right) }{\delta } \] (4) whenever \( f\left( {x,\alpha }\right) \) is sufficiently small and \( x \) is sufficiently near \( {x}_{0} \) . In cases where \( f \) is continuous, we then derive (4) for all \( \left( {x,\alpha }\right) \) sufficiently near \( \left( {{x}_{0},{\alpha }_{0}}\right) \) . Of course, part of the interest of (4) is that it implies the nonempti-ness of \( \Phi \left( \alpha \right) \) (since \( d\left( {x,\Phi \left( \alpha \right) }\right) = + \infty \) when \( \Phi \left( \alpha \right) \) is empty). The following application of the Solvability Theorem illustrates these remarks, together with the reduction of vector-valued equations to a single scalar one. We consider the equation \[ F\left( {x,\alpha }\right) = 0, \] (5) where \( F \) is a mapping from \( X \times M \) to \( Y \), with \( M \) a metric space and \( Y \) another Hilbert space. We assume that some open set \( U \) of \( X \) exists such that \( F \) is continuous on \( U \times M \), and such that the partial derivative \( {F}_{x}^{\prime }\left( {x,\alpha }\right) \) exists for all \( \left( {x,\alpha }\right) \in U \times M \), and is continuous there (jointly in \( \left( {x,\alpha }\right) \) ). 3.2. Theorem (Graves-Lyusternik). Let \( \left( {{x}_{0},{\alpha }_{0}}\right) \in U \times M \) be a point satisfying \( F\left( {{x}_{0},{\alpha }_{0}}\right) = 0 \), and suppose that \( {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) \) is onto: \[ {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) X = Y \] Let \( \Omega \) be any neighborhood of \( {x}_{0} \) . Then for some \( \delta > 0 \), for all \( \left( {x,\alpha }\right) \) sufficiently near \( \left( {{x}_{0},{\alpha }_{0}}\right) \), we have \[ d\left( {x;\Phi \left( \alpha \right) }\right) \leq \frac{\parallel F\left( {x,\alpha }\right) \parallel }{\delta } \] where \( \Phi \left( \alpha \right) \mathrel{\text{:=}} \left\{ {{x}^{\prime } \in \Omega : F\left( {{x}^{\prime },\alpha }\right) = 0}\right\} \) . Proof. By the Open Mapping Theorem of Banach, there exists \( \delta > 0 \) such that \( {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) {\bar{B}}_{X} \supset {2\delta }{\bar{B}}_{Y} \) . Since \( \left( {x,\alpha }\right) \mapsto {F}_{x}^{\prime }\left( {x,\alpha }\right) \) is continuous, there exist neighborhoods \( V \) and \( A \) of \( {x}_{0} \) and \( {\alpha }_{0} \) such that the operator norm of \( {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) - {F}_{x}^{\prime }\left( {x,\alpha }\right) \) is no greater than \( \delta \) when \( \left( {x,\alpha }\right) \in V \times A \) . For such \( \left( {x,\alpha }\right) \) we have \[ {2\delta }{\bar{B}}_{Y} \subset {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) {\bar{B}}_{X} \subset {F}_{x}^{\prime }\left( {x,\alpha }\right) {\bar{B}}_{X} + \left\lbrack {{F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) - {F}_{x}^{\prime }\left( {x,\alpha }\right) }\right\rbrack {\bar{B}}_{X} \] \[ \subset {F}_{x}^{\prime }\left( {x,\alpha }\right) {\bar{B}}_{X} + \delta {\bar{B}}_{Y} \] This implies that \[ {F}_{x}^{\prime }\left( {x,\alpha }\right) {\bar{B}}_{X} \supset \delta {\bar{B}}_{Y}\forall \left( {x,\alpha }\right) \in V \times A. \] It follows that for any unit vector \( y \) in \( Y \), for any \( \left( {x,\alpha }\right) \in V \times A \), we have \( \begin{Vmatrix}{{F}_{x}^{\prime }{\left( x,\alpha \right) }^{ * }y}\end{Vmatrix} \geq \delta \) . Indeed, there exists a vector \( v \in {\bar{B}}_{X} \) such that \( {F}_{x}^{\prime }\left( {x,\alpha }\right) v = {\delta y} \) . Then \[ \begin{Vmatrix}{{F}_{x}^{\prime }{\left( x,\alpha \right) }^{ * }y}\end{Vmatrix} \geq \left\langle {{F}_{x}^{\prime }{\left( x,\alpha \right) }^{ * }y, v}\right\rangle = \left\langle {y,{F}_{x}^{\prime }\left( {x,\alpha }\right) v}\right\rangle = \langle y,{\
1012_(GTM178)Nonsmooth Analysis and Control Theory
45
prime }\left( {{x}_{0},{\alpha }_{0}}\right) - {F}_{x}^{\prime }\left( {x,\alpha }\right) }\right\rbrack {\bar{B}}_{X} \] \[ \subset {F}_{x}^{\prime }\left( {x,\alpha }\right) {\bar{B}}_{X} + \delta {\bar{B}}_{Y} \] This implies that \[ {F}_{x}^{\prime }\left( {x,\alpha }\right) {\bar{B}}_{X} \supset \delta {\bar{B}}_{Y}\forall \left( {x,\alpha }\right) \in V \times A. \] It follows that for any unit vector \( y \) in \( Y \), for any \( \left( {x,\alpha }\right) \in V \times A \), we have \( \begin{Vmatrix}{{F}_{x}^{\prime }{\left( x,\alpha \right) }^{ * }y}\end{Vmatrix} \geq \delta \) . Indeed, there exists a vector \( v \in {\bar{B}}_{X} \) such that \( {F}_{x}^{\prime }\left( {x,\alpha }\right) v = {\delta y} \) . Then \[ \begin{Vmatrix}{{F}_{x}^{\prime }{\left( x,\alpha \right) }^{ * }y}\end{Vmatrix} \geq \left\langle {{F}_{x}^{\prime }{\left( x,\alpha \right) }^{ * }y, v}\right\rangle = \left\langle {y,{F}_{x}^{\prime }\left( {x,\alpha }\right) v}\right\rangle = \langle y,{\delta y}\rangle = \delta , \] as claimed. We are now ready to apply the Solvability Theorem 3.1, with \( f\left( {x,\alpha }\right) \mathrel{\text{:=}} \) \( \parallel F\left( {x,\alpha }\right) \parallel \) . We need to verify the nonstationarity condition of the theorem. If \( f\left( {x,\alpha }\right) > 0 \), then the only possible proximal subgradient of \( f\left( {\cdot ,\alpha }\right) \) at \( x \) is its derivative there, namely \( {F}_{x}^{\prime }{\left( x,\alpha \right) }^{ * }F\left( {x,\alpha }\right) /\parallel F\left( {x,\alpha }\right) \parallel \) (by the classical Chain Rule). We have shown above that such a vector has norm at least \( \delta \) , as required. Once the Solvability Theorem has been applied, there remains only to observe that (3) reduces to the required estimate for all \( \left( {x,\alpha }\right) \) sufficiently near \( \left( {{x}_{0},{\alpha }_{0}}\right) \), since then we will have \[ \min \{ d\left( {x;\operatorname{comp}V}\right), d\left( {x;\operatorname{comp}\Omega }\right) \} > \frac{f\left( {x,\alpha }\right) }{\delta }. \] We remark that the conclusion of the theorem subsumes as special cases stability results with respect to either a parameter change or to a change in the variable \( x \) . If \( \alpha \) alone changes (somewhat), then \( {x}_{0} \) is no longer a solution to \( F\left( {x,\alpha }\right) = 0 \), but is not far from one; if only \( x \) changes, then there is a solution for the original equation which approximates \( x \) to the same extent to which \( x \) is infeasible (i.e., \( \begin{Vmatrix}{F\left( {x,{\alpha }_{0}}\right) }\end{Vmatrix} \) ). (We had noted the first of these two properties in a different context in Theorem 1.9.) The term \( \parallel F\left( {x,\alpha }\right) \parallel \) which appears in the estimate given by Theorem 3.2 is nondifferentiable at \( \left( {{x}_{0},{\alpha }_{0}}\right) \), and this is necessarily the case: No term of the form \( \theta \left( {\parallel F\left( {x,\alpha }\right) \parallel }\right) \) with \( \theta \left( 0\right) = 0 \) and \( \theta \) differentiable can provide the estimate, as simple examples show. ## A Mixed Equality/Constraint System The following result will be used in the next theorem, which will illustrate the use of extended-valued functions \( f \) in connection with solvability, as well as the use of a constraint qualification. This refers to a nondegeneracy hypothesis on the data incorporating the constraints of the problem, similar to the normality hypothesis introduced in \( §{3.1} \) . 3.3. Lemma. Let \( F \) and \( \left( {{x}_{0},{\alpha }_{0}}\right) \) satisfy all the hypotheses of Theorem 3.2, where we take \( X \) and \( Y \) finite dimensional. Let \( S \) be a closed subset of \( X \) containing \( {x}_{0} \), and assume in addition that the following constraint qualification holds: \[ 0 \in {F}_{x}^{\prime }{\left( {x}_{0},{\alpha }_{0}\right) }^{ * }y + {N}_{S}^{L}\left( {x}_{0}\right) \Rightarrow y = 0. \] (6) Then there exist \( \delta > 0 \) and neighborhoods \( V \) and \( A \) of \( {x}_{0} \) and \( {\alpha }_{0} \) with the following property: for all \( \left( {x,\alpha }\right) \in V \times A \), for any unit vector \( y \), any vector \( \zeta \) belonging to \[ {F}_{x}^{\prime }{\left( x,\alpha \right) }^{ * }y + {N}_{S}^{L}\left( x\right) \] satisfies \( \parallel \zeta \parallel \geq \delta \) . Proof. If the assertion is false, then there exist sequences \( \left\{ {x}_{i}\right\} \) converging to \( {x}_{0},\left\{ {\alpha }_{i}\right\} \) converging to \( {\alpha }_{0},\left\{ {y}_{i}\right\} \) unit vectors such that, for some sequence \( \left\{ {\zeta }_{i}\right\} \) converging to 0, we have \[ {\zeta }_{i} \in {F}_{x}^{\prime }{\left( {x}_{i},{\alpha }_{i}\right) }^{ * }{y}_{i} + {N}_{S}^{L}\left( {x}_{i}\right) \] We can assume that the sequence \( {y}_{i} \) converges to a limit \( {y}_{0} \), necessarily a unit vector too. Since the multifunction \( x \mapsto {N}_{S}^{L}\left( x\right) \) has closed graph \( (X \) being finite dimensional), we obtain in the limit \( 0 \in {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) y + {N}_{S}^{L}\left( {x}_{0}\right) \) , contradicting the hypothesis (6). 3.4. Theorem. Let \( \left( {{x}_{0},{\alpha }_{0}}\right) \) be a solution of the equality/constraint system \[ F\left( {x,\alpha }\right) = 0,\;x \in S, \] under the hypotheses of Lemma 3.3, and in particular in the presence of the constraint qualification (6). Then for some \( \delta > 0 \), for all \( \left( {x,\alpha }\right) \) sufficiently near \( \left( {{x}_{0},{\alpha }_{0}}\right) \) having \( x \in S \), we have \[ d\left( {x;{\Phi }_{S}\left( \alpha \right) }\right) \leq \frac{\parallel F\left( {x,\alpha }\right) \parallel }{\delta } \] where \( {\Phi }_{S}\left( \alpha \right) \mathrel{\text{:=}} \{ x \in S : F\left( {x,\alpha }\right) = 0\} \) . In particular, \( {\Phi }_{S}\left( \alpha \right) \) is nonempty for \( \alpha \) sufficiently near \( {\alpha }_{0} \) . 3.5. Exercise. (a) Prove Theorem 3.4 with the help of Lemma 3.3, setting \[ f\left( {x,\alpha }\right) \mathrel{\text{:=}} \parallel F\left( {x,\alpha }\right) \parallel + {I}_{S}\left( x\right) \] in the Solvability Theorem and recalling Proposition 1.10.1. (b) When \( {x}_{0} \in \operatorname{int}S \) in the context of Lemma 3.3, show that (6) holds iff the matrix \( {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) \) has maximal rank. (c) A given function \( F : {\mathbb{R}}^{3} \times M \rightarrow {\mathbb{R}}^{2} \) satisfies the hypotheses of Theorem 3.2, with \[ {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) = \left\lbrack \begin{matrix} 1 & 0 & 0 \\ 1 & 1 & - 1 \end{matrix}\right\rbrack . \] Prove that for every \( \alpha \) sufficiently near \( {\alpha }_{0} \) there exists \( x = \) \( \left( {{x}_{1},{x}_{2},{x}_{3}}\right) \) in \( {\mathbb{R}}^{3} \) with \( {x}_{2} \geq 0,{x}_{3} \geq 0 \) such that \( F\left( {x,\alpha }\right) = 0 \) . In fact, show that we can also require that \( {x}_{2}{x}_{3} = 0 \) . ## Implicit and Inverse Functions Under additional structural assumptions, the equation \( F\left( {x,\alpha }\right) = 0 \) can be shown to uniquely define \( x \) as a function of \( \alpha \) (locally); the implicit function inherits regularity from \( F \) . We make the same assumptions as in Theorem 3.2, but now with \( X = Y \) . ## 3.6. Theorem. (a) Let \( {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) \) be onto and one-to-one. Then there exist neighborhoods \( \Omega \) of \( {x}_{0} \) and \( W \) of \( {\alpha }_{0} \) and a unique continuous function \( \widehat{x}\left( \cdot \right) : W \rightarrow \Omega \) with \( \widehat{x}\left( {\alpha }_{0}\right) = {x}_{0} \) such that \( F\left( {\widehat{x}\left( \alpha \right) ,\alpha }\right) = 0\forall \alpha \in W \) . (b) If in addition \( F \) is Lipschitz in a neighborhood of \( \left( {{x}_{0},{\alpha }_{0}}\right) \), then \( \widehat{x} \) is Lipschitz. (c) If in addition \( M \) is a Hilbert space and \( F\left( {x,\alpha }\right) \) is \( {C}^{1} \) (in both variables) near \( \left( {{x}_{0},{\alpha }_{0}}\right) \), then \( \widehat{x} \) is \( {C}^{1} \) . Proof. It is a well-known consequence of the Closed Graph Theorem that \( {F}_{x}^{\prime }{\left( {x}_{0},{\alpha }_{0}\right) }^{-1} \) is a continuous linear operator, and this can be used to show that \( {F}_{x}^{\prime }{\left( {x}_{0},{\alpha }_{0}\right) }^{ * } \) is onto (with inverse \( {\left( {F}_{x}^{\prime }{\left( {x}_{0},{\alpha }_{0}\right) }^{-1}\right) }^{ * } \) ). Thus for some \( \eta > \) 0, we have \( {F}_{x}^{\prime }{\left( {x}_{0},{\alpha }_{0}\right) }^{ * }\bar{B} \supset {2\eta }\bar{B} \) . For some convex neighborhood \( \Omega \times N \subset \) \( U \times M \) of \( \left( {{x}_{0},{\alpha }_{0}}\right) \), we have \[ \begin{Vmatrix}{{F}_{x}^{\prime }\left( {x,\alpha }\right) - {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) }\end{Vmatrix} < \eta \forall \left( {x,\alpha }\right) \in \Omega \times N. \] Now let \( x,{x}^{\prime } \in \Omega \) and \( \alpha \in N,{x}^{\prime } \neq x \) . Then, for any vector \( \theta \) in \( \bar{B} \) we have, by the Mean Value Theorem, for some \( z \in \left\lbrack {x,{x}^{\prime }}\right\rbrack \) : \[ \left\langle {\theta, F\left( {{x}^{\prime },\alpha }\right) - F\left( {x,\alpha }\right) }\right\rangle = \left\langle {\theta ,{F}_{x}^{\prime }\left( {z,\alpha }\right) \left( {{x}^{\prime } - x}\right) }\right\rangle \] \[ \geq \left\langle {\theta ,{F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) \left( {{x}^{\prime } - x}\right) }\right\rangle \] \[ + \left\langle {\theta ,\left( {{F}_{x}^{\prime }\left( {z,\alpha }\right) - {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) }\right) \left( {{x}^{\prime } - x}\right) }\right\rangle \] \[ \geq \left\langle {{F}_{x}^{\prime }{\left(
1012_(GTM178)Nonsmooth Analysis and Control Theory
46
}\right) - {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) }\end{Vmatrix} < \eta \forall \left( {x,\alpha }\right) \in \Omega \times N. \] Now let \( x,{x}^{\prime } \in \Omega \) and \( \alpha \in N,{x}^{\prime } \neq x \) . Then, for any vector \( \theta \) in \( \bar{B} \) we have, by the Mean Value Theorem, for some \( z \in \left\lbrack {x,{x}^{\prime }}\right\rbrack \) : \[ \left\langle {\theta, F\left( {{x}^{\prime },\alpha }\right) - F\left( {x,\alpha }\right) }\right\rangle = \left\langle {\theta ,{F}_{x}^{\prime }\left( {z,\alpha }\right) \left( {{x}^{\prime } - x}\right) }\right\rangle \] \[ \geq \left\langle {\theta ,{F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) \left( {{x}^{\prime } - x}\right) }\right\rangle \] \[ + \left\langle {\theta ,\left( {{F}_{x}^{\prime }\left( {z,\alpha }\right) - {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) }\right) \left( {{x}^{\prime } - x}\right) }\right\rangle \] \[ \geq \left\langle {{F}_{x}^{\prime }{\left( {x}_{0},{\alpha }_{0}\right) }^{ * }\theta ,{x}^{\prime } - x}\right\rangle - \eta \begin{Vmatrix}{{x}^{\prime } - x}\end{Vmatrix} \] \[ = \eta \begin{Vmatrix}{{x}^{\prime } - x}\end{Vmatrix} \] where \( \theta \) has been chosen in \( \bar{B} \) to satisfy \[ {F}_{x}^{\prime }{\left( {x}_{0},{\alpha }_{0}\right) }^{ * }\theta = {2\eta }\frac{\left( {x}^{\prime } - x\right) }{\begin{Vmatrix}{x}^{\prime } - x\end{Vmatrix}}. \] It follows that \[ \begin{Vmatrix}{F\left( {{x}^{\prime },\alpha }\right) - F\left( {x,\alpha }\right) }\end{Vmatrix} \geq \eta \begin{Vmatrix}{{x}^{\prime } - x}\end{Vmatrix}\forall \left( {x,\alpha }\right) \in \Omega \times N. \] (7) We now invoke Theorem 3.2 to deduce the existence of a neighborhood \( W \subset N \) of \( {\alpha }_{0} \) such that \[ \Phi \left( \alpha \right) \mathrel{\text{:=}} \{ x \in \Omega : F\left( {x,\alpha }\right) = 0\} \neq \varnothing \;\forall \alpha \in W. \] It follows from (7) that \( \Phi \left( \alpha \right) \) is a singleton \( \{ \widehat{x}\left( \alpha \right) \} \) for each \( \alpha \in W \), and that \( \widehat{x} \) is the unique function from \( W \) to \( \Omega \) such that \( F\left( {\widehat{x}\left( \alpha \right) ,\alpha }\right) = 0\forall \alpha \in W \) . We now show that \( \widehat{x}\left( \cdot \right) \) is continuous. From (7) we have \[ \eta \begin{Vmatrix}{\widehat{x}\left( {\alpha }^{\prime }\right) - \widehat{x}\left( \alpha \right) }\end{Vmatrix} \leq \begin{Vmatrix}{F\left( {\widehat{x}\left( {\alpha }^{\prime }\right) ,{\alpha }^{\prime }}\right) - F\left( {\widehat{x}\left( \alpha \right) ,{\alpha }^{\prime }}\right) }\end{Vmatrix} = \begin{Vmatrix}{F\left( {\widehat{x}\left( \alpha \right) ,{\alpha }^{\prime }}\right) }\end{Vmatrix}, \] which tends to 0 as \( {\alpha }^{\prime } \) tends to \( \alpha \) . This establishes continuity. In addition, if \( F \) is Lipschitz near \( \left( {{x}_{0},{\alpha }_{0}}\right) \), then for \( \alpha \) and \( {\alpha }^{\prime } \) sufficiently near \( {\alpha }_{0} \) the last term above satisfies \[ \begin{Vmatrix}{F\left( {\widehat{x}\left( \alpha \right) ,{\alpha }^{\prime }}\right) }\end{Vmatrix} = \begin{Vmatrix}{F\left( {\widehat{x}\left( \alpha \right) ,{\alpha }^{\prime }}\right) - F\left( {\widehat{x}\left( \alpha \right) ,\alpha }\right) }\end{Vmatrix} \leq K{d}_{M}\left( {\alpha ,{\alpha }^{\prime }}\right) , \] for some suitable Lipschitz constant \( K \), where \( {d}_{M} \) is the metric on \( M \) . We deduce that \( \widehat{x}\left( \cdot \right) \) is Lipschitz near \( {\alpha }_{0} \), proving (b). Finally, let \( M \) be a Hilbert space and let \( F \) be \( {C}^{1} \) in \( \left( {x,\alpha }\right) \) . The Taylor expansion gives \[ 0 = F\left( {\widehat{x}\left( \alpha \right) ,\alpha }\right) = F\left( {{x}_{0},{\alpha }_{0}}\right) + {F}_{x}^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) \left( {\widehat{x}\left( \alpha \right) - {x}_{0}}\right) + {F}_{\alpha }^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) \left( {\alpha - {\alpha }_{0}}\right) \] \[ + o\left( {\widehat{x}\left( \alpha \right) - {x}_{0},\alpha - {\alpha }_{0}}\right) \] where \[ o\left( {x - {x}_{0},\alpha - {\alpha }_{0}}\right) /\begin{Vmatrix}\left( {x - {x}_{0},\alpha - {\alpha }_{0}}\right) \end{Vmatrix} \rightarrow 0\text{ as }\left( {x,\alpha }\right) \rightarrow \left( {{x}_{0},{\alpha }_{0}}\right) . \] Since \( \widehat{x}\left( \cdot \right) \) is Lipschitz near \( {\alpha }_{0}, o\left( {\widehat{x}\left( \alpha \right) - {x}_{0},\alpha - {\alpha }_{0}}\right) \) can be written \( o\left( {\alpha - {\alpha }_{0}}\right) \) . Then we can solve the Taylor expansion explicitly for \( \widehat{x}\left( \alpha \right) \) and deduce from the resulting expression that \( \widehat{x}\left( \cdot \right) \) is differentiable at \( {\alpha }_{0} \), with derivative \( - {F}_{x}^{\prime }{\left( {x}_{0},{\alpha }_{0}\right) }^{-1}{F}_{\alpha }^{\prime }\left( {{x}_{0},{\alpha }_{0}}\right) \) . But this argument holds not just at \( {\alpha }_{0} \) but at any \( \alpha \) in a neighborhood, whence \( \widehat{x}\left( \cdot \right) \) is \( {C}^{1} \) near \( {\alpha }_{0} \) . A special case of the theorem is that of inverse functions: 3.7. Corollary. Let \( F \) satisfy the hypotheses of the theorem and have the form \( F\left( {x,\alpha }\right) = G\left( x\right) - \alpha \) (so that \( X = Y = M \) ). Then there exist neighborhoods \( W \) of \( {\alpha }_{0} \) and \( \Lambda \) of \( {x}_{0} \) and a \( {C}^{1} \) function \( \widehat{x} \) on \( W \) such that \[ G\left( {\widehat{x}\left( \alpha \right) }\right) = \alpha \forall \alpha \in W,\;\widehat{x}\left( {G\left( x\right) }\right) = x\forall x \in \Lambda . \] Proof. The function \( \widehat{x}\left( \cdot \right) \) of the theorem is \( {C}^{1} \) and satisfies \( G\left( {\widehat{x}\left( \alpha \right) }\right) = \alpha \) for \( \alpha \in W \), the neighborhood of \( {\alpha }_{0} \) provided by the theorem. Now pick \( \Lambda \subset \Omega \) so that \( G\left( x\right) \in W\forall x \in \Lambda \) . Then, as shown in the theorem, \( \Phi \left( {G\left( x\right) }\right) \) is a singleton when \( x \in \Lambda \) ; that is, there is a unique \( {x}^{\prime } \in \Omega \) such that \( G\left( {x}^{\prime }\right) = G\left( x\right) \) . But then this \( {x}^{\prime } \), namely \( \widehat{x}\left( {G\left( x\right) }\right) \), must equal \( x \) . ## Systems of Mixed Equalities and Inequalities We consider now the solutions of a parametrized system of the form \[ h\left( x\right) = \alpha ,\;g\left( x\right) \leq \beta , \] (8) where \( x \in X \) (a Hilbert space), and where the vector \( \left( {\alpha ,\beta }\right) \in {\mathbb{R}}^{m} \times {\mathbb{R}}^{p} \) is the parameter. Let \( {x}_{0} \) satisfy the system for \( \left( {\alpha ,\beta }\right) = \left( {0,0}\right) \), and let us suppose that the functions \( g : X \rightarrow {\mathbb{R}}^{p} \) and \( h : X \rightarrow {\mathbb{R}}^{m} \) are Lipschitz in a neighborhood \( \Omega \) of \( {x}_{0} \) . We set \[ \Phi \left( {\alpha ,\beta }\right) \mathrel{\text{:=}} \{ x \in \Omega : \text{ system }\left( 8\right) \text{ is satisfied }\} . \] 3.8. Theorem. Suppose that the following constraint qualification is satisfied at \( {x}_{0} \) : \[ \gamma \geq 0,\;\left\langle {\gamma, g\left( {x}_{0}\right) }\right\rangle = 0,\;0 \in {\partial }_{L}\{ \langle \gamma, g\left( \cdot \right) \rangle + \langle \lambda, h\left( \cdot \right) \rangle \} \left( {x}_{0}\right) \] \[ \Rightarrow \gamma = 0\;\text{ and }\;\lambda = 0. \] Then for some \( \delta > 0 \), for all \( x \) sufficiently near \( {x}_{0} \), and for all \( \left( {\alpha ,\beta }\right) \) sufficiently near \( \left( {0,0}\right) \), we have \[ d\left( {x;\Phi \left( {\alpha ,\beta }\right) }\right) \leq \frac{1}{\delta }\mathop{\max }\limits_{{i, j}}\left\{ {{\left( {g}_{i}\left( x\right) - {\beta }_{i}\right) }_{ + },\left| {{h}_{j}\left( x\right) - {\alpha }_{j}}\right| }\right\} \] where \( {\left( {g}_{i}\left( x\right) - {\beta }_{i}\right) }_{ + } \) signifies \( \max \left( {{g}_{i}\left( x\right) - {\beta }_{i},0}\right) \) . In particular, \( \Phi \left( {\alpha ,\beta }\right) \) is nonempty for \( \left( {\alpha ,\beta }\right) \) near \( \left( {0,0}\right) \) . Proof. We set \[ f\left( {x,\alpha ,\beta }\right) \mathrel{\text{:=}} \mathop{\max }\limits_{{i, j}}\left\{ {{\left( {g}_{i}\left( x\right) - {\beta }_{i}\right) }_{ + },\left| {{h}_{j}\left( x\right) - {\alpha }_{j}}\right| }\right\} \geq 0. \] Then \( x \) is a solution of (8) iff \( f\left( {x,\alpha ,\beta }\right) = 0 \) . Note that \( f \) is Lipschitz in a neighborhood of \( \left( {{x}_{0},0,0}\right) \), as observed in Problem 1.11.17. Lemma. There exist neighborhoods \( V \) of \( {x}_{0} \) and \( A \) of \( \left( {0,0}\right) \) and \( \delta > 0 \) such that \[ x \in V,\;\left( {\alpha ,\beta }\right) \in A,\;f\left( {x,\alpha ,\beta }\right) > 0,\;\zeta \in {\partial }_{P}f\left( {x,\alpha ,\beta }\right) \Rightarrow \parallel \zeta \parallel \geq \delta . \] Proof. If this fails to be the case, then there exist sequences \( \left\{ {x}_{k}\right\} ,\left\{ \left( {{\alpha }_{k},{\beta }_{k}}\right) \right\} \) , \( \left\{ {\zeta }_{k}\right\} \) such that \[ {x}_{k} \rightarrow {x}_{0},\;\left( {{\alpha }_{k},{\beta }_{k}}\right) \rightarrow \left( {0,0}\right) ,\;f\left( {{x}_{k},{\alpha }_{k},{\beta }_{k}}\right) > 0, \] \[ {\zeta }_{k} \in {\partial }_{P}f\left( {{x}_{k},{\alpha }_{k},{\beta }_{k}}\right) ,\;\begin{Vmatrix}{\zeta }_{k}\end{Vmatrix} \rightarrow 0. \] The proximal subgradient \( {\zeta }_{k} \) can be characterized via Problem 1.11.17, for \( k \) large enough so that \( {x}_{k} \) enters some prescribed neighborhood on which \( f \) is Lipschitz in \( x \) . Note that in the formula provided there, only the terms \( {\left( {g}_{i}\left( x\r
1012_(GTM178)Nonsmooth Analysis and Control Theory
47
,\beta }\right) \in A,\;f\left( {x,\alpha ,\beta }\right) > 0,\;\zeta \in {\partial }_{P}f\left( {x,\alpha ,\beta }\right) \Rightarrow \parallel \zeta \parallel \geq \delta . \] Proof. If this fails to be the case, then there exist sequences \( \left\{ {x}_{k}\right\} ,\left\{ \left( {{\alpha }_{k},{\beta }_{k}}\right) \right\} \) , \( \left\{ {\zeta }_{k}\right\} \) such that \[ {x}_{k} \rightarrow {x}_{0},\;\left( {{\alpha }_{k},{\beta }_{k}}\right) \rightarrow \left( {0,0}\right) ,\;f\left( {{x}_{k},{\alpha }_{k},{\beta }_{k}}\right) > 0, \] \[ {\zeta }_{k} \in {\partial }_{P}f\left( {{x}_{k},{\alpha }_{k},{\beta }_{k}}\right) ,\;\begin{Vmatrix}{\zeta }_{k}\end{Vmatrix} \rightarrow 0. \] The proximal subgradient \( {\zeta }_{k} \) can be characterized via Problem 1.11.17, for \( k \) large enough so that \( {x}_{k} \) enters some prescribed neighborhood on which \( f \) is Lipschitz in \( x \) . Note that in the formula provided there, only the terms \( {\left( {g}_{i}\left( x\right) - {\beta }_{i}\right) }_{ + } \) and \( \left| {{h}_{j}\left( x\right) - {\alpha }_{j}}\right| \) which are strictly positive can contribute. Locally, these coincide with \( {g}_{i}\left( x\right) - {\beta }_{i} \) and \( \pm \left( {{h}_{j}\left( x\right) - {\alpha }_{j}}\right) \) . The upshot is that for all \( k \) large enough, \( {\zeta }_{k} \) belongs to the set \[ {\partial }_{L}\left\{ {\left\langle {{\gamma }^{k}, g\left( \cdot \right) }\right\rangle + \left\langle {{\lambda }^{k}, h\left( \cdot \right) }\right\rangle }\right\} \left( {x}_{k}\right) \] where \[ {\gamma }_{i}^{k} \geq 0,\;\left\langle {{\gamma }^{k}, g\left( {x}_{k}\right) - {\beta }_{k}}\right\rangle = 0,\;\mathop{\sum }\limits_{i}{\gamma }_{i}^{k} + \mathop{\sum }\limits_{j}\left| {\lambda }_{j}^{k}\right| = 1. \] If we now extract subsequences as required to make each sequence \( \left\{ {\gamma }^{k}\right\} \) , \( \left\{ {\lambda }^{k}\right\} \) converge, passage to the limit (see Exercise 1.6) produces a nonzero \( \left( {\gamma ,\lambda }\right) \) with precisely the properties ruled out by the constraint qualification. This contradiction proves the lemma. Armed with the lemma, the application of the Solvability Theorem 3.1 is immediate and gives the required result. When the inequalities are absent in (8), it is natural to seek nonsmooth analogues of the implicit function assertions of Theorem 3.6. We will examine for illustrative purposes the issue of an inverse function induced by an equation \( G\left( x\right) = \alpha \), which was treated in the case of smooth data in Corollary 3.7. The solvability aspect is at hand, as a special case of the preceding theorem: 3.9. Exercise. Let \( G : X \rightarrow {\mathbb{R}}^{n} \) be Lipschitz on a neighborhood of the point \( {x}_{0} \), where \( G\left( {x}_{0}\right) = {\alpha }_{0} \) . Suppose that the following constraint qualification holds: \[ 0 \in {\partial }_{L}\langle \zeta, G\left( \cdot \right) \rangle \left( {x}_{0}\right) \Rightarrow \zeta = 0. \] (9) Then, for any neighborhood \( \Omega \) of \( {x}_{0} \), there exists \( \delta > 0 \) such that for all \( \left( {x,\alpha }\right) \) sufficiently near \( \left( {{x}_{0},{\alpha }_{0}}\right) \) we have \[ d\left( {x;\Phi \left( \alpha \right) }\right) \leq \frac{\parallel G\left( x\right) - \alpha \parallel }{\delta } \] where \( \Phi \left( \alpha \right) \mathrel{\text{:=}} \{ x \in \Omega : G\left( x\right) = \alpha \} \) . Show how this reduces to a special case of Theorem 3.2 when \( G \) is \( {C}^{1} \) . ## Generalized Jacobians and Inverse Functions What is required now to deduce the existence of an inverse is a condition akin to (7) which implies that \( G \) is locally one-to-one. A convenient way to formulate such a condition, which will also turn out to be a more readily verifiable criterion implying the constraint qualification (9), is in terms of the generalized Jacobian \( \partial G \), a tool which is applicable when the underlying space is finite dimensional. Let \( X = {\mathbb{R}}^{m} \) . We define \( \partial G\left( x\right) \) as follows: \[ \partial G\left( x\right) \mathrel{\text{:=}} \operatorname{co}\left\{ {\mathop{\lim }\limits_{{i \rightarrow \infty }}{G}^{\prime }\left( {x}_{i}\right) : {G}^{\prime }\left( {x}_{i}\right) \text{ exists,}{x}_{i} \rightarrow x}\right\} , \] a formula that brings to mind the Generalized Gradient Formula of Theorem 2.8.1. (We remark that, in keeping with the earlier formula, an arbitrary set of measure 0 can be avoided in calculating \( \partial G\left( x\right) \), without the results below being affected.) We identify \( \partial G\left( x\right) \) with a convex set of \( n \times m \) matrices, and we say that \( \partial G\left( x\right) \) is of maximal rank provided that every matrix in \( \partial G\left( x\right) \) is of maximal rank. When \( n = m \), it is alternate terminology to say that \( \partial G\left( x\right) \) is nonsingular. ## 3.10. Exercise. (a) Let \( G : {\mathbb{R}}^{m} \rightarrow {\mathbb{R}}^{n} \) be Lipschitz near \( x \) . Prove that \( \partial G\left( x\right) \) is compact, convex, and nonempty, where the space of \( n \times m \) matrices is identified with \( {\mathbb{R}}^{nm} \) . Prove that \( \partial G \) is upper semicontinuous: For all \( \varepsilon > 0 \) there exists \( r > 0 \) such that \[ {x}^{\prime } \in x + {rB} \Rightarrow \partial G\left( {x}^{\prime }\right) \subset \partial G\left( x\right) + \varepsilon {B}_{n \times m}, \] where \( {B}_{n \times m} \) denotes the open unit ball in the space of \( n \times m \) matrices. (b) Show that for any \( \zeta \in {\mathbb{R}}^{n} \), we have \( \partial \langle \zeta, G\left( \cdot \right) \rangle \left( x\right) = {\zeta }^{ * }\partial G\left( x\right) \) . (c) Calculate \( \partial G\left( 0\right) \) where \( n = m = 2 \) and \[ G\left( {x, y}\right) = \left( {\left| x\right| + y,{2x} + \left| y\right| }\right) \] and show that it is nonsingular. (d) If \( G \) has component functions \( {g}_{1},{g}_{2},\ldots ,{g}_{n} \), show that \[ \partial G\left( x\right) \subset \partial {g}_{1}\left( x\right) \times \partial {g}_{2}\left( x\right) \times \cdots \times \partial {g}_{n}\left( x\right) , \] and give an example in which strict containment occurs. 3.11. Proposition. Let \( G : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{n} \) be Lipschitz near \( {x}_{0} \), and suppose that \( \partial G\left( {x}_{0}\right) \) is nonsingular. Then the constraint qualification (9) holds, and for some \( \eta > 0 \), for some neighborhood \( \Omega \) of \( {x}_{0} \), we have \[ \parallel G\left( x\right) - G\left( y\right) \parallel \geq \eta \parallel x - y\parallel \forall x, y \in \Omega . \] Proof. Let 0 belong to \( {\partial }_{L}\langle \zeta, G\left( \cdot \right) \rangle \left( {x}_{0}\right) \) . Then \[ 0 \in \partial \langle \zeta, G\left( \cdot \right) \rangle \left( {x}_{0}\right) = {\zeta }^{ * }\partial G\left( {x}_{0}\right) \] (by Exercise 3.10). Since \( \partial G\left( {x}_{0}\right) \) is nonsingular, we deduce \( \zeta = 0 \), thereby verifying the constraint qualification. The next step in the proof requires the following: Lemma. There are positive numbers \( r \) and \( \eta \) with the following property. Given any unit vector \( v \in {\mathbb{R}}^{n} \), there is a unit vector \( w \) in \( {\mathbb{R}}^{n} \) such that, whenever \( x \) lies in \( {x}_{0} + {rB} \) and the matrix \( M \) belongs to \( \partial G\left( x\right) \), then we have \[ \langle w,{Mv}\rangle \geq \eta \] Proof. To see this, let \( S \) be the unit sphere in \( {\mathbb{R}}^{n} \) and note that the subset \( \partial G\left( {x}_{0}\right) S \) of \( {\mathbb{R}}^{n} \) is compact and does not contain 0 . Hence, \[ d\left( {0;\partial G\left( {x}_{0}\right) S}\right) /2 = : \eta > 0. \] For some \( \varepsilon > 0 \) small enough, we will have \[ d\left( {0;\left( {\partial G\left( {x}_{0}\right) + \varepsilon {B}_{n \times n}}\right) S}\right) > \eta . \] Now for some \( r > 0 \), for all \( x \in {x}_{0} + {rB} \), we have \( \partial G\left( x\right) \subset \partial G\left( {x}_{0}\right) + \varepsilon {B}_{n \times n} \) , in light of Exercise 3.10, where \( r \) is small enough so that \( G \) is Lipschitz on \( {x}_{0} + r\bar{B} \) . Now let any unit vector \( v \) be given. It follows from the above that the convex set \( \left( {\partial G\left( {x}_{0}\right) + \varepsilon {B}_{n \times n}}\right) v \) is distance at least \( \eta \) from 0 . By the Separation Theorem, there exists a unit vector \( w \) such that \( \langle \zeta, w\rangle \geq \eta \) for all \( \zeta \in \left( {\partial G\left( {x}_{0}\right) + \varepsilon {B}_{n \times n}}\right) v \), and all the more for any \( \zeta \) of the form \( {Mv} \) , \( M \in \partial G\left( x\right), x \in {x}_{0} + {rB} \) . This establishes the lemma. Now let \( x \) and \( y \) belong to \( \Omega \mathrel{\text{:=}} {x}_{0} + {rB} \) . We will complete the proof of the proposition by showing that \( \parallel G\left( x\right) - G\left( y\right) \parallel \geq \eta \parallel x - y\parallel \) . We may suppose \( x \neq y \) . Set \[ v = \frac{y - x}{\parallel y - x\parallel },\;\lambda = \parallel y - x\parallel \] so that \( y = x + {\lambda v} \) . Let \( \pi \) be the plane perpendicular to \( v \) and passing through \( x \) . The set \( P \) of points \( {x}^{\prime } \) in \( {x}_{0} + {rB} \) where \( {G}^{\prime }\left( {x}^{\prime }\right) \) fails to exist is of measure 0, and hence by Fubini’s Theorem, for almost every \( {x}^{\prime } \) in \( \pi \), the ray \[ {x}^{\prime } + {tv},\;t \geq 0 \] meets \( P \) in a set of 0 one-dimensional measure. Choose an \( {x}^{\prime } \) with the above property and sufficiently close to \( x \) so that \( {x}^{\prime } + {tv} \) lies in \( {x}_{0} + {rB} \) for every \( t \) in \( \left\lbrack {0,\la
1012_(GTM178)Nonsmooth Analysis and Control Theory
48
) belong to \( \Omega \mathrel{\text{:=}} {x}_{0} + {rB} \) . We will complete the proof of the proposition by showing that \( \parallel G\left( x\right) - G\left( y\right) \parallel \geq \eta \parallel x - y\parallel \) . We may suppose \( x \neq y \) . Set \[ v = \frac{y - x}{\parallel y - x\parallel },\;\lambda = \parallel y - x\parallel \] so that \( y = x + {\lambda v} \) . Let \( \pi \) be the plane perpendicular to \( v \) and passing through \( x \) . The set \( P \) of points \( {x}^{\prime } \) in \( {x}_{0} + {rB} \) where \( {G}^{\prime }\left( {x}^{\prime }\right) \) fails to exist is of measure 0, and hence by Fubini’s Theorem, for almost every \( {x}^{\prime } \) in \( \pi \), the ray \[ {x}^{\prime } + {tv},\;t \geq 0 \] meets \( P \) in a set of 0 one-dimensional measure. Choose an \( {x}^{\prime } \) with the above property and sufficiently close to \( x \) so that \( {x}^{\prime } + {tv} \) lies in \( {x}_{0} + {rB} \) for every \( t \) in \( \left\lbrack {0,\lambda }\right\rbrack \) . Then the function \[ t \rightarrow G\left( {{x}^{\prime } + {tv}}\right) \] is Lipschitzian for \( t \) in \( \left\lbrack {0,\lambda }\right\rbrack \) and has almost everywhere on this interval derivative \( {G}^{\prime }\left( {{x}^{\prime } + {tv}}\right) v \) . Thus \[ G\left( {{x}^{\prime } + {\lambda v}}\right) - G\left( {x}^{\prime }\right) = {\int }_{0}^{\lambda }{G}^{\prime }\left( {{x}^{\prime } + {tv}}\right) {vdt}. \] Let \( w \) correspond to \( v \) in the lemma above. We deduce \[ w \cdot \left\lbrack {G\left( {{x}^{\prime } + {\lambda v}}\right) - G\left( {x}^{\prime }\right) }\right\rbrack = {\int }_{0}^{\lambda }w \cdot \left\lbrack {{G}^{\prime }\left( {{x}^{\prime } + {tv}}\right) v}\right\rbrack {dt} \geq {\int }_{0}^{\lambda }{\eta dt} = {\lambda \eta }. \] Recalling the definition of \( \lambda \), we arrive at: \[ \begin{Vmatrix}{G\left( {{x}^{\prime } + {\lambda v}}\right) - G\left( {x}^{\prime }\right) }\end{Vmatrix} \geq \eta \parallel x - y\parallel . \] This conclusion has been obtained for almost all \( {x}^{\prime } \) in a neighborhood of \( x \) . Since \( G \) is continuous, the proposition is proved. We now possess all the necessary components to prove the Lipschitz Inverse Function Theorem. 3.12. Theorem. If \( G : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{n} \) is Lipschitz near \( {x}_{0} \) and \( \partial G\left( {x}_{0}\right) \) is nonsingular, then there exist neighborhoods \( W \) of \( G\left( {x}_{0}\right) \) and \( \Lambda \) of \( {x}_{0} \) and a Lipschitz function \( \widehat{x} \) on \( W \) such that \[ G\left( {\widehat{x}\left( \alpha \right) }\right) = \alpha \forall \alpha \in W,\;\widehat{x}\left( {G\left( x\right) }\right) = x\forall x \in \Lambda . \] 3. Special Topics ## 3.13. Exercise. (a) Prove Theorem 3.12. (b) Show that the theorem fails (even for \( n = 1 \) ) if the hypothesis that \( \partial G\left( {x}_{0}\right) \) be nonsingular is replaced by the condition that \( {G}^{\prime }\left( x\right) \) be nonsingular whenever it exists. (c) Confirm the conclusion of the theorem for the example of Exercise 3.10(c). ## 4 Derivate Calculus and Rademacher's Theorem The first constructs of nonsmooth analysis were put forward by Dini in the nineteenth century. His well-known derivates were defined for a real-valued function of a single variable, but the basic idea can be extended beyond that setting and developed so as to include corresponding notions of subdifferential, tangent, and normal. We will take a brief look at this theory now, and relate it to the results of the two preceding chapters. The setting will be that of a function \( f \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) ; for \( x \in \operatorname{dom}f \), we define the subderivate of \( f \) at \( x \) in the direction \( v \), denoted \( {Df}\left( {x;v}\right) \), as follows: \[ {Df}\left( {x;v}\right) \mathrel{\text{:=}} \mathop{\liminf }\limits_{\substack{{w \rightarrow v} \\ {t \downarrow 0} }}\frac{f\left( {x + {tw}}\right) - f\left( x\right) }{t}. \] 4.1. Exercise. (a) If \( f \) is Lipschitz of rank \( K \) near \( x \), then \( {Df}\left( {x;v}\right) \) agrees with \[ \mathop{\liminf }\limits_{{t \downarrow 0}}\frac{f\left( {x + {tv}}\right) - f\left( x\right) }{t} \] and we have \( {Df}\left( {x;0}\right) = 0 \) . The function \( v \mapsto {Df}\left( {x;v}\right) \) is Lipschitz of rank \( K \) on \( {\mathbb{R}}^{n} \), and \( {Df}\left( {x;v}\right) \leq {f}^{ \circ }\left( {x;v}\right) \) . Equality holds for all \( v \) iff \( f \) is regular at \( x \) . (b) In general, \( {Df}\left( {x;0}\right) \) is either 0 or \( - \infty \), and we have \( {Df}\left( {x;{\lambda v}}\right) = \) \( {\lambda Df}\left( {x;v}\right) \) for any \( \lambda > 0 \) . Give an example of a continuous \( f : \mathbb{R} \rightarrow \mathbb{R} \) having \( {Df}\left( {0;0}\right) = - \infty \) . (c) If \( g \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) has \( x \in \operatorname{dom}g \), then \[ D\left( {f + g}\right) \left( {x;v}\right) \geq {Df}\left( {x;v}\right) + {Dg}\left( {x;v}\right) \] (interpreting \( \infty - \infty \) as \( - \infty \) ). (d) If \( \zeta \in {\partial }_{P}f\left( x\right) \), then \( {Df}\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v \) . (e) The function \( v \mapsto {Df}\left( {x;v}\right) \) is lower semicontinuous. The following theorem is the key to relating derivate calculus to proximal analysis. 4.2. Theorem (Subbotin). Let \( f \in \mathcal{F}\left( {\mathbb{R}}^{n}\right), x \in \operatorname{dom}f \), and let \( E \) be a nonempty compact convex subset of \( {\mathbb{R}}^{n} \) . Suppose that for some scalar \( \rho \) we have \[ {Df}\left( {x;e}\right) > \rho \forall e \in E. \] Then, for any \( \varepsilon > 0 \), there exist \( z \in x + {\varepsilon B} \) and \( \zeta \in {\partial }_{P}f\left( z\right) \) such that \[ \left| {f\left( z\right) - f\left( x\right) }\right| < \varepsilon ,\;\langle \zeta, e\rangle > \rho \forall e \in E. \] Proof. We claim that for all \( t > 0 \) sufficiently small, we have \[ f\left( {x + {te} + {t}^{2}u}\right) - f\left( x\right) > {\rho t} + {t}^{2}\forall e \in E,\forall u \in \bar{B}. \] (1) Were this not so, there would be sequences \( {t}_{i} \downarrow 0,{e}_{i} \in E \), and \( {u}_{i} \in \bar{B} \) such that \[ \frac{f\left( {x + {t}_{i}{e}_{i} + {t}_{i}^{2}{u}_{i}}\right) - f\left( x\right) }{{t}_{i}} \leq \rho + {t}_{i} \] Invoking the compactness of \( E \), there is a subsequence (we eschew relabeling) of \( {e}_{i} \) converging to a point \( {e}_{0} \in E \) . Then \( {e}_{i} + {t}_{i}{u}_{i} \) converges to \( {e}_{0} \) as well, and we deduce \( {Df}\left( {x;{e}_{0}}\right) \leq \rho \), contradicting the hypothesis and establishing the claim. Now pick \( t > 0 \) so that (1) holds, so that \( {tE} + {t}^{2}B \subset {\varepsilon B} \), so that \( t < \varepsilon /\left( {2\rho }\right) \) , and finally also small enough so that \( f\left( {x}^{\prime }\right) > f\left( x\right) - \varepsilon \) for all \( {x}^{\prime } \in x + {tE} + {t}^{2}\bar{B} \) (this is possible because \( f \) is lower semicontinuous). We will proceed to apply Theorem 2.6, the Mean Value Inequality, for \( Y \mathrel{\text{:=}} x + {tE} \) . We remark that in light of (1), the \( \widehat{r} \) of the theorem satisfies \( \widehat{r} \geq {\rho t} + {t}^{2} \), so \( \bar{r} = {\rho t} \) is a suitable choice in applying the theorem. We pick for the \( \varepsilon \) of the theorem any positive \( {\varepsilon }^{\prime } < \min \left\lbrack {\varepsilon /2,{t}^{2}}\right\rbrack \) . The point \( z \) provided by the theorem lies in \( \left\lbrack {x, Y}\right\rbrack + {\varepsilon }^{\prime }B \), which gives \[ z \in x + {tE} + {t}^{2}B \subset x + {\varepsilon B}, \] whence \( f\left( z\right) > f\left( x\right) - \varepsilon \) by choice of \( t \) . Also, Theorem 2.6 asserts \[ f\left( z\right) < f\left( x\right) + \bar{r} + {\varepsilon }^{\prime } < f\left( x\right) + {\rho t} + \frac{\varepsilon }{2} < f\left( x\right) + \varepsilon . \] Thus \( \left| {f\left( z\right) - f\left( x\right) }\right| < \varepsilon \) as desired. Finally we have, for the designated \( \zeta \) in \( {\partial }_{P}f\left( z\right) \) \[ \langle \zeta ,{te}\rangle > \bar{r} \mathrel{\text{:=}} {\rho t}\forall e \in E, \] which is the other assertion of the theorem. 4.3. Exercise. Show that the theorem is false if the convexity of \( E \) is deleted from the hypotheses, by considering \( X = {\mathbb{R}}^{2}, f\left( u\right) \mathrel{\text{:=}} \parallel u\parallel \) , \( x = 0, E = \) unit circle. ## The \( D \) -Subdifferential The classical formula \( {f}^{\prime }\left( {x;v}\right) = \langle \nabla f\left( x\right), v\rangle \) encapsulates the duality between derivate (directional derivative) constructions and differential ones. Motivated by this, as well as by the duality between \( {f}^{ \circ }\left( {x; \cdot }\right) \) and \( \partial f\left( x\right) \) observed in Chapter 2, we proceed to define a subdifferential that corresponds to the subderivate \( {Df}\left( {x; \cdot }\right) \) . We say that \( \zeta \) is a directional subgradient or \( D \) -subgradient of \( f \) at \( x \) provided that \( x \in \operatorname{dom}f \) and \[ {Df}\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v \in {\mathbb{R}}^{n}. \] We denote the set of all such \( \zeta \) by \( {\partial }_{D}f\left( x\right) \) and refer to it as the \( D \) - subdifferential. It follows from Exercise 4.1(d) that \( {\partial }_{D}f\left( x\right) \) contains \( {\partial }_{P}f\left( x\right) \) , and so must be nonempty on a dense subset of \( \operatorname{dom}f \) by the Proximal Density Theorem 1.3.1. 4.4. Exercise. (a) Show that \( {\partial }_{D}f\left( x\right) \) is closed and convex, and reduces to \( \left\{ {{f}^{\prime }\left( x\right) }\right\} \) if \( f
1012_(GTM178)Nonsmooth Analysis and Control Theory
49
tween \( {f}^{ \circ }\left( {x; \cdot }\right) \) and \( \partial f\left( x\right) \) observed in Chapter 2, we proceed to define a subdifferential that corresponds to the subderivate \( {Df}\left( {x; \cdot }\right) \) . We say that \( \zeta \) is a directional subgradient or \( D \) -subgradient of \( f \) at \( x \) provided that \( x \in \operatorname{dom}f \) and \[ {Df}\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v \in {\mathbb{R}}^{n}. \] We denote the set of all such \( \zeta \) by \( {\partial }_{D}f\left( x\right) \) and refer to it as the \( D \) - subdifferential. It follows from Exercise 4.1(d) that \( {\partial }_{D}f\left( x\right) \) contains \( {\partial }_{P}f\left( x\right) \) , and so must be nonempty on a dense subset of \( \operatorname{dom}f \) by the Proximal Density Theorem 1.3.1. 4.4. Exercise. (a) Show that \( {\partial }_{D}f\left( x\right) \) is closed and convex, and reduces to \( \left\{ {{f}^{\prime }\left( x\right) }\right\} \) if \( f \) is (Fréchet) differentiable at \( x \) . (b) Prove that \( {\partial }_{D}f\left( x\right) \) is bounded if \( f \) is Lipschitz near \( x \) . (c) We have \[ {\partial }_{D}\left( {{f}_{1} + {f}_{2}}\right) \left( x\right) \supseteq {\partial }_{D}{f}_{1}\left( x\right) + {\partial }_{D}{f}_{2}\left( x\right) . \] Equality holds if one of the functions is differentiable at \( x \) . (d) For \( f\left( x\right) = - \parallel x\parallel \), show that \( {\partial }_{D}f\left( 0\right) = \varnothing \) . (e) Recall that for \( f\left( x\right) = - {\left| x\right| }^{3/2} \) we have \( {\partial }_{P}f\left( x\right) = \varnothing \) ; show that \( {\partial }_{D}f\left( 0\right) = \{ 0\} \) (f) There is no true duality between \( {Df}\left( {x; \cdot }\right) \) and \( {\partial }_{D}f\left( x\right) \) : It is not true in general that \( {Df}\left( {x;v}\right) = \sup \left\{ {\langle \zeta, v\rangle : \zeta \in {\partial }_{D}f\left( x\right) }\right\} \) . ## Approximation by \( P \) -Subgradients We now prove that although \( {\partial }_{D}f\left( x\right) \) can be strictly larger than \( {\partial }_{P}f\left( x\right) \) , the essential difference between the two in a local sense (as opposed to pointwise) is slight. 4.5. Proposition. Let \( \zeta \in {\partial }_{D}f\left( x\right) \) . Then for any \( \varepsilon > 0 \) there exist \( z \in \) \( x + {\varepsilon B} \) and \( \xi \in {\partial }_{P}f\left( z\right) \) such that \( \left| {f\left( x\right) - f\left( z\right) }\right| < \varepsilon \) and \( \parallel \zeta - \xi \parallel < \varepsilon \) . Proof. It follows from the definition of \( {\partial }_{D}f\left( x\right) \) that we have \[ {D\varphi }\left( {x;v}\right) \geq 0\forall v \in {\mathbb{R}}^{n}, \] where \[ \varphi \left( y\right) \mathrel{\text{:=}} f\left( y\right) - \langle \zeta, y\rangle \] Then for arbitrary fixed \( \delta > 0 \) , \[ {D\varphi }\left( {x;v}\right) > - \delta \forall v \in \bar{B}. \] We now apply Subbotin's Theorem 4.2 in order to obtain the existence of \( z \) and \( {\xi }^{\prime } \in {\partial }_{P}\varphi \left( z\right) \) such that \[ z \in x + {\delta B},\;\left| {\varphi \left( z\right) - \varphi \left( x\right) }\right| < \delta , \] and \[ \left\langle {{\xi }^{\prime }, v}\right\rangle > - \delta \forall v \in \bar{B} \] (2) It follows that \[ \left| {f\left( z\right) - f\left( x\right) }\right| \leq \left| {\varphi \left( z\right) - \varphi \left( x\right) }\right| + \parallel \zeta \parallel \parallel x - z\parallel < \left( {1 + \parallel \zeta \parallel }\right) \delta , \] (3) and from (2), that \( \begin{Vmatrix}{\xi }^{\prime }\end{Vmatrix} < \delta \) . But \[ {\partial }_{P}\varphi \left( z\right) = {\partial }_{P}f\left( z\right) - \zeta \] which implies the existence of \( \xi \in {\partial }_{P}f\left( z\right) \) such that \( {\xi }^{\prime } = \xi - \zeta \) . Hence, bearing (3) in mind, if we choose \[ \delta < \frac{\varepsilon }{1 + \parallel \zeta \parallel } \] then \( z \) and \( \xi \) fulfill the requirements of the proposition. 4.6. Exercise. Let \( {f}^{\prime }\left( x\right) \) exist. Then for any \( \varepsilon > 0 \) there exist \( z \in \) \( x + {\varepsilon B} \) and \( \zeta \in {\partial }_{P}f\left( z\right) \) such that \( \left| {f\left( x\right) - f\left( z\right) }\right| < \varepsilon \) and \( \begin{Vmatrix}{{f}^{\prime }\left( x\right) - \zeta }\end{Vmatrix} < \) \( \varepsilon \) . ## Directional Calculus We have seen in the preceding proposition that the graph of the multifunction \( {\partial }_{P}f \) is dense in that of \( {\partial }_{D}f \) . This fact can be used to derive rather easily the basic calculus of \( {\partial }_{D}f \) by appealing to the known results for \( {\partial }_{P}f \) . Here are a few examples, the first of which goes back to a result of Dini that can be viewed as the first theorem of nonsmooth analysis. 4.7. Exercise. (a) Let \( f \in \mathcal{F}\left( \mathbb{R}\right) \) . Then the following are equivalent: (i) \( f \) is decreasing; (ii) \( {Df}\left( {x;1}\right) \leq 0\forall x \) ; and (iii) \( \zeta \leq 0\forall \zeta \in {\partial }_{D}f\left( x\right) \forall x \) . (Compare with Theorem 0.1.1.) (b) Let \( {f}_{1},{f}_{2} \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \), and let \( \zeta \in {\partial }_{D}\left( {{f}_{1} + {f}_{2}}\right) \left( x\right) \) . Then for any \( \varepsilon > \) 0 there exist points \( {x}_{1} \) and \( {x}_{2} \) in \( x + {\varepsilon B} \) with \( \left| {{f}_{i}\left( x\right) - {f}_{i}\left( {x}_{i}\right) }\right| < \varepsilon \) \( \left( {i = 1,2}\right) \) such that \[ \zeta \in {\partial }_{D}{f}_{1}\left( {x}_{1}\right) + {\partial }_{D}{f}_{2}\left( {x}_{2}\right) + {\varepsilon B}. \] (c) A function \( f \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) is Lipschitz of rank \( K \) near \( x \in \operatorname{dom}f \) iff there is a neighborhood \( U \) of \( x \) such that \[ {Df}\left( {{x}^{\prime };v}\right) \leq K\parallel v\parallel \forall v \in {\mathbb{R}}^{n},\forall {x}^{\prime } \in U. \] The fact that \( D \) -subgradients can be approximated by proximal ones implies that \( D \) -subgradients generate the same limiting subdifferential \( {\partial }_{L}f\left( x\right) \) as that studied in \( §{1.10} \), and also has a bearing on the generalized gradient \( \partial f\left( x\right) \) and regularity (§2.4), as we now see. We remind the reader that \( {\partial }_{C}f \) is an alternative notation for the generalized gradient (see Remark 2.6.2). 4.8. Proposition. Let \( f \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) and \( x \in \operatorname{dom}f \) . (a) \( {\partial }_{L}f\left( x\right) = \left\{ {\mathop{\lim }\limits_{{i \rightarrow \infty }}{\zeta }_{i} : {\zeta }_{i} \in {\partial }_{D}f\left( {x}_{i}\right) ,{x}_{i}\overset{f}{ \rightarrow }x}\right\} \) . (b) If \( f \) is Lipschitz near \( x \), then \[ {\partial }_{D}f\left( x\right) \subset {\partial }_{L}f\left( x\right) \subset {\partial }_{C}f\left( x\right) \] with equality throughout iff \( f \) is regular at \( x \) (in particular, when \( f \) is convex). Proof. Part (a) is immediate from Proposition 4.5, and implies the first inclusion in (b); the second has been noted earlier in connection with Theorem 2.6.1. Suppose now that \( f \) is regular at \( x \), and let \( \zeta \in {\partial }_{C}f\left( x\right) \) . Then, for any \( v \in {\mathbb{R}}^{n} \) , \[ \langle \zeta, v\rangle \leq {f}^{ \circ }\left( {x;v}\right) \;\text{ (by definition of }{\partial }_{C}f\left( x\right) \text{ ) } \] \[ = {f}^{\prime }\left( {x;v}\right) \;\text{ (by definition of regularity) } \] \[ = {Df}\left( {x;v}\right) \;\text{(in light of Exercise 4.1(a)).} \] Thus \( \zeta \in {\partial }_{D}f\left( x\right) \), and we have \( {\partial }_{D}f\left( x\right) = {\partial }_{L}f\left( x\right) = {\partial }_{C}f\left( x\right) \) . For the converse, observe that the equality of the three sets in question implies that \( {Df}\left( {x;v}\right) = {f}^{ \circ }\left( {x;v}\right) \) for all \( v \), since then \[ {Df}\left( {x;v}\right) \geq \max \left\{ {\langle \zeta, v\rangle : \zeta \in {\partial }_{D}f\left( x\right) }\right\} \] \[ = \max \left\{ {\langle \zeta, v\rangle : \zeta \in {\partial }_{C}f\left( x\right) }\right\} = {f}^{ \circ }\left( {x;v}\right) \] and since the opposite inequality always holds. That this implies regularity at \( x \) was noted in Exercise 4.1. ## Tangents and Normals A natural way to define a tangent direction \( v \) to a set \( S \) at \( x \in S \) is to require \( D{d}_{S}\left( {x;v}\right) = 0 \) (or equivalently, \( \leq 0 \) ), where \( {d}_{S} \) is the distance function associated with \( S \) . The set of such \( v \) constitutes the \( D \) -tangent cone \( {T}_{S}^{D}\left( x\right) \) . Taking a now familiar path, we define the \( D \) -normal cone by polarity: \( {N}_{S}^{D}\left( x\right) \mathrel{\text{:=}} {T}_{S}^{D}{\left( x\right) }^{ \circ } \) . The following exercise includes the geometric counterpart of Proposition 4.8. 4.9. Exercise. Let \( S \) be a nonempty closed subset of \( {\mathbb{R}}^{n} \) . (a) \( {T}_{S}^{D}\left( x\right) \) coincides with the Bouligand tangent cone \( {T}_{S}^{B}\left( x\right) \) of §2.7. (b) We do not have full duality: In general, \( {T}_{S}^{D}\left( x\right) \) is not the polar of \( {N}_{S}^{D}\left( x\right) \) . (c) \( \zeta \in {N}_{S}^{D}\left( x\right) \) iff \[ \mathop{\limsup }\limits_{{{x}^{\prime }\overset{S}{ \rightarrow }x}}\frac{\left\langle \zeta ,{x}^{\prime } - x\right\rangle }{\begin{Vmatrix}{x}^{\prime } - x\end{Vmatrix}} \leq 0. \] (4) (d) Any unit vector \( \zeta \) satisfying (4) belongs to \( {\partial }_{D}{d}_{S}\left( x\right) \) . (e) \( {N}_{S}^{D}\left( x\right) \) is the cone generated by \( {\partial }_{D}{d}_{S}\left( x\right) \) (compare with Theorem 1.6.4, Propositi
1012_(GTM178)Nonsmooth Analysis and Control Theory
50
^{D}\left( x\right) \mathrel{\text{:=}} {T}_{S}^{D}{\left( x\right) }^{ \circ } \) . The following exercise includes the geometric counterpart of Proposition 4.8. 4.9. Exercise. Let \( S \) be a nonempty closed subset of \( {\mathbb{R}}^{n} \) . (a) \( {T}_{S}^{D}\left( x\right) \) coincides with the Bouligand tangent cone \( {T}_{S}^{B}\left( x\right) \) of §2.7. (b) We do not have full duality: In general, \( {T}_{S}^{D}\left( x\right) \) is not the polar of \( {N}_{S}^{D}\left( x\right) \) . (c) \( \zeta \in {N}_{S}^{D}\left( x\right) \) iff \[ \mathop{\limsup }\limits_{{{x}^{\prime }\overset{S}{ \rightarrow }x}}\frac{\left\langle \zeta ,{x}^{\prime } - x\right\rangle }{\begin{Vmatrix}{x}^{\prime } - x\end{Vmatrix}} \leq 0. \] (4) (d) Any unit vector \( \zeta \) satisfying (4) belongs to \( {\partial }_{D}{d}_{S}\left( x\right) \) . (e) \( {N}_{S}^{D}\left( x\right) \) is the cone generated by \( {\partial }_{D}{d}_{S}\left( x\right) \) (compare with Theorem 1.6.4, Proposition 2.5.4, and Problem 1.11.27.) (f) \( {N}_{S}^{D}\left( x\right) \subset {N}_{S}^{L}\left( x\right) \subset {N}_{S}^{C}\left( x\right) \), with equality throughout iff \( S \) is regular at \( x \) . (Hint. To prove the first inclusion, use part (e), approximate \( {\partial }_{D}{d}_{S} \) by \( {\partial }_{P}{d}_{S} \), and recall Theorems 2.6.1 and 1.6.4.) ## Other Characterizations of \( {\partial }_{D}f \) An alternate characterization of \( {\partial }_{D}f \) that bypasses \( {Df} \) is provided by the following: 4.10. Proposition. \( \zeta \in {\partial }_{D}f\left( x\right) \) iff \[ \mathop{\liminf }\limits_{\substack{{u \rightarrow 0} \\ {u \neq 0} }}\frac{f\left( {x + u}\right) - f\left( x\right) -\langle \zeta, u\rangle }{\parallel u\parallel } \geq 0. \] (5) Proof. Let (5) hold, and let \( \left\{ {v}_{i}\right\} \) be a sequence converging to \( v \) and \( \left\{ {t}_{i}\right\} \) a sequence decreasing to 0 . Then (5) implies \[ \mathop{\liminf }\limits_{{i \rightarrow \infty }}\frac{f\left( {x + {t}_{i}{v}_{i}}\right) - f\left( x\right) }{{t}_{i}} \geq \langle \zeta, v\rangle \] whence \( {Df}\left( {x;v}\right) \geq \langle \zeta, v\rangle \forall v \) . Thus \( \zeta \in {\partial }_{D}f\left( x\right) \) . For the converse, let \( \zeta \in {\partial }_{D}f\left( x\right) \) and suppose that (5) fails. Then there is a sequence \( \left\{ {u}_{i}\right\} \) and \( \varepsilon > 0 \) such that \( {u}_{i} \rightarrow 0 \) and \[ f\left( {x + {u}_{i}}\right) - f\left( x\right) \leq \left\langle {\zeta ,{u}_{i}}\right\rangle - \varepsilon \begin{Vmatrix}{u}_{i}\end{Vmatrix}. \] Note that \( {u}_{i} \neq 0 \) . We can extract a subsequence (without relabeling) so that the vectors \( {v}_{i} \mathrel{\text{:=}} {u}_{i}/\begin{Vmatrix}{u}_{i}\end{Vmatrix} \) converge to a limit \( v \) . Setting \( {t}_{i} \mathrel{\text{:=}} \begin{Vmatrix}{u}_{i}\end{Vmatrix} \), the last inequality becomes \[ \frac{f\left( {x + {t}_{i}{v}_{i}}\right) - f\left( x\right) }{{t}_{i}} \leq \left\langle {\zeta ,{v}_{i}}\right\rangle - \varepsilon \] which implies \[ {Df}\left( {x;v}\right) \leq \langle \zeta, v\rangle - \varepsilon \] contradicting \( \zeta \in {\partial }_{D}f\left( x\right) \) and completing the proof. We remark that the set of subgradients defined via (5) is sometimes referred to as the Fréchet subdifferential \( {\partial }_{F}f\left( x\right) \), an object which can differ from \( {\partial }_{D}f\left( x\right) \) in infinite dimensions. ## 4.11. Exercise. (a) Show that (5) holds iff there exists a nonnegative function \( o\left( r\right) \) such that \( o\left( r\right) /r \rightarrow 0 \) as \( r \downarrow 0 \) and \[ f\left( {x + u}\right) - f\left( x\right) + o\left( {\parallel u\parallel }\right) \geq \langle \zeta, u\rangle \] for all \( u \) near 0 . Show that \( o\left( \cdot \right) \) can always be taken to be increasing. By considering \[ \bar{o}\left( t\right) \mathrel{\text{:=}} t\max \{ o\left( r\right) /r : 0 < r \leq t\} \] show that \( o \) can also be assumed to have the property that \( o\left( t\right) /t \) is increasing. (b) \( \zeta \in {\partial }_{D}f\left( x\right) \) iff for every \( \varepsilon > 0 \) there exists a neighborhood \( {N}_{\varepsilon } \) of \( x \) such that \( x \) minimizes the function \[ {g}_{\varepsilon }\left( y\right) \mathrel{\text{:=}} f\left( y\right) - f\left( x\right) - \langle \zeta, y - x\rangle + \varepsilon \parallel y - x\parallel \] over \( {N}_{\varepsilon } \) . (c) Show that the function \( o \) of part (a) can also be assumed to be continuous. (Hint. Show that the following function has all the required properties: \[ \widehat{o}\left( t\right) \mathrel{\text{:=}} {2t}{\int }_{t}^{2t}\frac{o\left( r\right) }{{r}^{2}}{dr} \] The following is another, less intrinsic but interesting characterization of \( {\partial }_{D}f \) . It is the definition of subdifferential most often put forward in the literature on viscosity solutions of differential equations, and \( {\partial }_{D}f \) is sometimes referred to as the "viscosity subdifferential." 4.12. Proposition. Let \( f \in \mathcal{F} \) . Then \( \zeta \in {\partial }_{D}f\left( x\right) \) iff there exists a continuous function \( g : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) which is differentiable at \( x \) with \( {g}^{\prime }\left( x\right) = \zeta \), and such that \( f - g \) has a local minimum at \( x \) . Proof. First suppose that there exists \( g \) with the properties mentioned. Then \( 0 \in {\partial }_{D}\left( {f - g}\right) \left( x\right) \), and Exercise 4.4(c) implies \( {g}^{\prime }\left( x\right) \in {\partial }_{D}f\left( x\right) \) . Now suppose that \( \zeta \in {\partial }_{D}f\left( x\right) \) . Consider the (ultimate) function \( o\left( \cdot \right) \) in Exercise 4.11(a). Since \( o\left( \cdot \right) \) is increasing, it is continuous except at countably many points. Since \( o\left( r\right) /r \downarrow 0 \) as \( r \downarrow 0 \), this function is integrable on any bounded interval, and we may therefore define a function \( \varphi : \lbrack 0,\infty ) \rightarrow \) \( ( - \infty ,0\rbrack \) as follows: \[ \varphi \left( r\right) \mathrel{\text{:=}} - {\int }_{r}^{2r}\frac{o\left( s\right) }{s}{ds}. \] It is readily noted that \( - o\left( r\right) \geq \varphi \left( r\right) \geq - o\left( {2r}\right) /2 \) for all positive \( r \), whence \( \varphi \left( r\right) \rightarrow 0 \) as \( r \downarrow 0 \) . We also have \[ f\left( y\right) \geq g\left( y\right) \mathrel{\text{:=}} f\left( x\right) + \langle \zeta, y - x\rangle + \varphi \left( {\parallel y - x\parallel }\right) , \] for all \( y \) near \( x \) . Since \( \varphi \left( 0\right) = 0 \), it follows that \( f - g \) has a local minimum at \( x \) . Clearly \( g \) is continuous, so it remains only to verify that \( {g}^{\prime }\left( x\right) \) exists and equals \( \zeta \) ; i.e., that the function \( w\left( y\right) \mathrel{\text{:=}} \varphi \left( {\parallel y - x\parallel }\right) \) has \( {w}^{\prime }\left( x\right) = 0 \) . For any nonzero \( v \) in \( {\mathbb{R}}^{n} \) we have \[ 0 \geq \frac{w\left( {x + {tv}}\right) - w\left( x\right) }{t} = \frac{\varphi \left( {\parallel {tv}\parallel }\right) }{t} = \parallel v\parallel \frac{\varphi \left( {\parallel {tv}\parallel }\right) }{\parallel {tv}\parallel } \geq \frac{-o\left( {{2t}\parallel v\parallel }\right) }{{2t}\parallel v\parallel }\left( {\parallel v\parallel }\right) . \] Upon taking the limit as \( t \downarrow 0 \) we see that the convergence to 0 is uniform on bounded sets of \( v \), and so the Fréchet derivative of \( w \) at \( x \) exists and satisfies \( {w}^{\prime }\left( x\right) = 0 \) . This completes the proof. 4.13. Exercise. Show that Proposition 4.12 remains true if in its statement the word "continuous" is replaced by "Lipschitz" or by "continuously differentiable." What if it is replaced by " \( {C}^{2} \) "? Of course all the results of this section concerning \( {Df} \) and \( {\partial }_{D}f \) have counterparts in terms of upper derivates and supergradients of an upper semicontinuous function \( f \) . When \( f \) is continuous, we can simply define the \( D \) -superdifferential \( {\partial }^{D}f\left( x\right) \) to be \( - {\partial }_{D}\left( {-f}\right) \left( x\right) \) . 4.14. Exercise. Let \( f \) be continuous near \( x \), and suppose that both \( {\partial }_{D}f\left( x\right) \) and \( {\partial }^{D}f\left( x\right) \) are nonempty. Prove that \( {f}^{\prime }\left( x\right) \) exists, and that \[ {\partial }_{D}f\left( x\right) = {\partial }^{D}f\left( x\right) = \left\{ {{f}^{\prime }\left( x\right) }\right\} . \] ## Partial Subdifferentials Let \( f\left( {x, y}\right) \) be a function of two real variables, and suppose that the partial derivatives \( \partial f/\partial x \) and \( \partial f/\partial y \) both exist at \( \left( {0,0}\right) \) . As we know from classical theory, it does not follow that \( f \) is differentiable at \( \left( {0,0}\right) \) ; some extra hypothesis is required to give this conclusion. (The classical result requires that one of the partials be defined and continuous in \( \left( {x, y}\right) \) near \( \left( {0,0}\right) \) .) In other words, differentiability cannot be confirmed "one direction at a time." Similar considerations hold in the nonsmooth setting: that \( f \) has a proximal subgradient \( \zeta \) relative to \( x \) and a proximal subgradient \( \xi \) relative to \( y \) does not imply that \( \left( {\zeta ,\xi }\right) \) is a proximal subgradient of \( f \) as a function of both variables (though the converse is true). In the setting of the \( D \) -calculus, the following exercise illustrates the same point. 4.15. Exercise. Let \( f\left( {x, y}\right) \) be a Lipschitz function on \( {\mathbb{R}}^{2} \) . (a) If \( \partial f/\partial x\left( {0,0}\right) \)
1012_(GTM178)Nonsmooth Analysis and Control Theory
51
\right) \) . As we know from classical theory, it does not follow that \( f \) is differentiable at \( \left( {0,0}\right) \) ; some extra hypothesis is required to give this conclusion. (The classical result requires that one of the partials be defined and continuous in \( \left( {x, y}\right) \) near \( \left( {0,0}\right) \) .) In other words, differentiability cannot be confirmed "one direction at a time." Similar considerations hold in the nonsmooth setting: that \( f \) has a proximal subgradient \( \zeta \) relative to \( x \) and a proximal subgradient \( \xi \) relative to \( y \) does not imply that \( \left( {\zeta ,\xi }\right) \) is a proximal subgradient of \( f \) as a function of both variables (though the converse is true). In the setting of the \( D \) -calculus, the following exercise illustrates the same point. 4.15. Exercise. Let \( f\left( {x, y}\right) \) be a Lipschitz function on \( {\mathbb{R}}^{2} \) . (a) If \( \partial f/\partial x\left( {0,0}\right) \) exists, show that \[ {\partial }_{D}\left( {f\left( {\cdot ,0}\right) }\right) \left( 0\right) = \{ \partial f/\partial x\left( {0,0}\right) \} \] where \( {\partial }_{D}\left( {f\left( {\cdot ,0}\right) }\right) \left( 0\right) \) signifies the \( D \) -subdifferential of the function \( g\left( x\right) \mathrel{\text{:=}} f\left( {x,0}\right) \) at the point \( x = 0 \) . If \( \left( {\zeta ,\xi }\right) \in {\partial }_{D}f\left( {0,0}\right) \) , show that \( \zeta \) belongs to \( {\partial }_{D}\left( {f\left( {\cdot ,0}\right) }\right) \left( 0\right) \) . Deduce that if the gradient \( \nabla f\left( {0,0}\right) \) exists (i.e., the vector whose components are the two partial derivatives of \( f \) at \( \left( {0,0}\right) ) \), then \[ {\partial }_{D}f\left( {0,0}\right) \subseteq \{ \nabla f\left( {0,0}\right) \} \] (b) By considering the case \( f\left( {x, y}\right) = - \min \left\lbrack {\left| x\right| ,\left| y\right| }\right\rbrack \), show that \( {\partial }_{D}f\left( {0,0}\right) \) can be empty even if \( \nabla f\left( {0,0}\right) \) exists. Despite the rather negative nature of this example, there does turn out to be a relation in finite dimensions between partial and full subdifferentials that holds "often" in the presence of Lipschitz behavior. The proof of the following will suppose known the fact that a locally Lipschitz function of a single variable is differentiable almost everywhere in the sense of Lebesgue measure on the line. 4.16. Theorem. Let \( f : {\mathbb{R}}^{n} \times \mathbb{R} \rightarrow \mathbb{R} \) be a Lipschitz function. Then for all \( x \in {\mathbb{R}}^{n} \), for \( y \) in \( \mathbb{R} \) almost everywhere, the following holds: \[ {\partial }_{D}f\left( {x, y}\right) = {\partial }_{D}f\left( {\cdot, y}\right) \left( x\right) \times {\partial }_{D}f\left( {x, \cdot }\right) \left( y\right) \] \[ = {\partial }_{D}f\left( {\cdot, y}\right) \left( x\right) \times \left\{ {\frac{\partial f}{\partial y}\left( {x, y}\right) }\right\} . \] (6) Proof. Fix any \( x \in {\mathbb{R}}^{n} \), and let us limit attention to the set \( Y \) of \( y \) for which \( \left( {\partial f/\partial y}\right) \left( {x, y}\right) \) exists, a set which differs from \( \mathbb{R} \) by a null set, and which is therefore measurable. The left side of (6) is always contained in the right, for any \( y \in Y \), by Exercise 4.15(a), so it suffices to prove that the opposite inclusion holds for almost all \( y \in Y \) . In turn, this would follow from the fact that for almost all \( y \in Y \), we have \[ {Df}\left( {x, y;v, w}\right) \geq {Df}\left( {x, y;v,0}\right) + w\frac{\partial f}{\partial y}\left( {x, y}\right) \forall v \in {\mathbb{R}}^{n},\forall w \in \mathbb{R}. \] (7) This is what we will prove. The object, then, is to prove that the following subset \( S \) of \( \mathbb{R} \) has measure 0 (or is contained in a set of measure zero, since Lebesgue measure is complete): \[ S \mathrel{\text{:=}} \left\{ {y \in Y : \exists \left( {v, w}\right) \text{ such that }{Df}\left( {x, y;v,0}\right) }\right. \] \[ \left. { > {Df}\left( {x, y;v, w}\right) - w\frac{\partial f}{\partial y}\left( {x, y}\right) }\right\} . \] Note that a direction \( \left( {v, w}\right) \) corresponding to \( y \) as in the definition of \( S \) must have \( w \neq 0 \) . If \( y \in S \) and \( \left( {v, w}\right) \) is (one of) its corresponding directions, then there exist \( r \in \mathbb{R} \) and \( \varepsilon > 0 \) such that \[ \frac{f\left( {x + {tv}, y}\right) - f\left( {x, y}\right) }{t} > r > {Df}\left( {x, y;v, w}\right) - w\frac{\partial f}{\partial y}\left( {x, y}\right) \] \[ \text{for}0 < t < \varepsilon \text{.} \] (8) We label \( C\left( {v, w, r,\varepsilon }\right) \) the set of \( y \in Y \) for which (8) holds. Now let \( \left\{ {v}_{i}\right\} ,\left\{ {w}_{i}\right\} ,\left\{ {r}_{i}\right\} \), and \( \left\{ {\varepsilon }_{i}\right\} \) be countable dense sets in \( {\mathbb{R}}^{n},\mathbb{R},\mathbb{R} \), and \( \left( {0,\infty }\right) \), respectively. Note that all the functions of \( \left( {v, w}\right) \) appearing in (8) are continuous (see Exercise 4.1(a)), and (crucial fact!) that \[ \left| {\frac{f\left( {x + {tv}, y}\right) }{t} - \frac{f\left( {x + t{v}_{i}, y}\right) }{t}}\right| \leq K\begin{Vmatrix}{v - {v}_{i}}\end{Vmatrix} \] independently of \( t > 0 \), where \( K \) is the Lipschitz constant for \( f \) . It follows that if (8) holds for a given \( y \), then it also holds (for the same \( y \) ) with \( \left( {v, w, r,\varepsilon }\right) \) replaced by suitably close \( \left( {{v}_{i},{w}_{i},{r}_{i},{\varepsilon }_{i}}\right) \) (with \( {\varepsilon }_{i} < \varepsilon \) ). In other words, \( S \) is contained in the countable union of the sets \( C\left( {{v}_{i},{w}_{i},{r}_{i},{\varepsilon }_{i}}\right) \), and so it suffices to prove that each such set has measure 0 . We will drop the indices for this last part of the proof, and we suppose \( w \neq 0 \) (for otherwise \( C\left( {v, w, r,\varepsilon }\right) \) is empty). We will need some measurability results. Lemma. The function \( y \mapsto {Df}\left( {x, y;v, w}\right) \) is measurable. Proof. For each positive integer \( j \), let \( {\left\{ {t}_{i}^{j}\right\} }_{i = 1}^{\infty } \) be a countable dense set in \( \left( {0,1/j}\right) \) . Then, for any \( \alpha \in \mathbb{R} \), we have \[ \{ y \in \mathbb{R} : {Df}\left( {x, y;v, w}\right) > \alpha \} \] \[ = \mathop{\bigcup }\limits_{j}\mathop{\bigcap }\limits_{i}\left\{ {y \in \mathbb{R} : \frac{f\left( {x + {t}_{i}^{j}v, y + {t}_{i}^{j}w}\right) - f\left( {x, y}\right) }{{t}_{i}^{j}} > \alpha + \frac{1}{j}}\right\} . \] The right side is a countable union/intersection of measurable sets, and so is itself measurable. The lemma follows. We now ask the reader to contribute the following facts to the proof: 4.17. Exercise. (a) The function \( y \rightarrow \partial f/\partial y\left( {x, y}\right) \) defined on \( Y \) is measurable. (b) The subset \( C\left( {v, w, r,\varepsilon }\right) \) of \( Y \) defined by (8) is measurable. Now let \( {y}_{0} \) belong to \( C\left( {v, w, r,\varepsilon }\right) \) ; i.e., let (8) hold. Choose \( \delta > 0 \) so that \( \delta < {4K}\left| w\right| \), and small enough so that \[ r > {Df}\left( {x,{y}_{0};v, w}\right) - w\frac{\partial f}{\partial y}\left( {x,{y}_{0}}\right) + \delta , \] (9) and let \( \left\{ {t}_{i}\right\} \) be a sequence "realizing" \( {Df}\left( {x,{y}_{0};v, w}\right) \) ; i.e., such that \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{f\left( {x + {t}_{i}v,{y}_{0} + {t}_{i}w}\right) - f\left( {x,{y}_{0}}\right) }{{t}_{i}} = {Df}\left( {x,{y}_{0};v, w}\right) . \] Of course, we also have \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{f\left( {x,{y}_{0} + {t}_{i}w}\right) - f\left( {x,{y}_{0}}\right) }{{t}_{i}} = w\frac{\partial f}{\partial y}\left( {x,{y}_{0}}\right) . \] It follows from (9) that for some \( {i}_{0} \), for all \( i \geq {i}_{0} \), the following holds: \[ r > \frac{f\left( {x + {t}_{i}v,{y}_{0} + {t}_{i}w}\right) - f\left( {x,{y}_{0}}\right) }{{t}_{i}} \] \[ - \frac{f\left( {x,{y}_{0} + {t}_{i}w}\right) - f\left( {x,{y}_{0}}\right) }{{t}_{i}} + \frac{\delta }{2} \] \[ = \frac{f\left( {x + {t}_{i}v,{y}_{0} + {t}_{i}w}\right) - f\left( {x,{y}_{0} + {t}_{i}w}\right) }{{t}_{i}} + \frac{\delta }{2}. \] (10) We now consider any point \( y \in C\left( {v, w, r,\varepsilon }\right) \) (i.e., satisfying (8)). Then, putting \( t = {t}_{i} \), we have, for \( i \) sufficiently large, \[ \frac{f\left( {x + {t}_{i}v, y}\right) - f\left( {x, y}\right) }{{t}_{i}} > r. \] Combining this with (10) for \( i \geq {i}_{0} \) and using the Lipschitz constant \( K \) for \( f \) yields \[ {2K}\frac{\left| y - {y}_{0} - {t}_{i}w\right| }{{t}_{i}} > \frac{\delta }{2} \] This says that for all \( i \) sufficiently large, we have \[ \left\{ {\left( {{y}_{0} + {t}_{i}w}\right) + \frac{\delta }{4K}{t}_{i}B}\right\} \cap S = \varnothing . \] Thus, letting \( {I}_{i} \) be the interval whose endpoints are \( {y}_{0} \) and \( {y}_{0} + {t}_{i}w \), and putting \( C \mathrel{\text{:=}} C\left( {v, w, r,\varepsilon }\right) \), we deduce (where \( \mathcal{L} \) denotes Lebesgue measure on \( \mathbb{R} \) ), \[ \mathcal{L}\left\{ {{I}_{i} \cap C}\right\} \leq \left( {1 - \frac{\delta }{{4K}\left| w\right| }}\right) {t}_{i}\left| w\right| ,\;i \geq {i}_{0}. \] Consequently, \[ \mathop{\limsup }\limits_{{i \rightarrow \infty }}\frac{\mathcal{L}\left( {{I}_{i} \cap C}\right) }{\mathcal{L}\left( {I}_{i}\right) } \leq 1 - \frac{\delta }{{4K}\left| w\right| } < 1. \] (11) Let \( g : \mathbb{R} \rightarrow \mathbb{R} \) be the function \[ g\left( t\right) \mathrel{\text{:=}} {\int }_{0}^{t}{\chi }_{C}\left( s\right) {ds} = \mathcal{L}\{ \left\lbrack {0, t}\right\rbrack \cap C\} \] where \( {\chi }_{C} \) is the c
1012_(GTM178)Nonsmooth Analysis and Control Theory
52
\[ \left\{ {\left( {{y}_{0} + {t}_{i}w}\right) + \frac{\delta }{4K}{t}_{i}B}\right\} \cap S = \varnothing . \] Thus, letting \( {I}_{i} \) be the interval whose endpoints are \( {y}_{0} \) and \( {y}_{0} + {t}_{i}w \), and putting \( C \mathrel{\text{:=}} C\left( {v, w, r,\varepsilon }\right) \), we deduce (where \( \mathcal{L} \) denotes Lebesgue measure on \( \mathbb{R} \) ), \[ \mathcal{L}\left\{ {{I}_{i} \cap C}\right\} \leq \left( {1 - \frac{\delta }{{4K}\left| w\right| }}\right) {t}_{i}\left| w\right| ,\;i \geq {i}_{0}. \] Consequently, \[ \mathop{\limsup }\limits_{{i \rightarrow \infty }}\frac{\mathcal{L}\left( {{I}_{i} \cap C}\right) }{\mathcal{L}\left( {I}_{i}\right) } \leq 1 - \frac{\delta }{{4K}\left| w\right| } < 1. \] (11) Let \( g : \mathbb{R} \rightarrow \mathbb{R} \) be the function \[ g\left( t\right) \mathrel{\text{:=}} {\int }_{0}^{t}{\chi }_{C}\left( s\right) {ds} = \mathcal{L}\{ \left\lbrack {0, t}\right\rbrack \cap C\} \] where \( {\chi }_{C} \) is the characteristic function of \( C \) . Then we know from integration theory on the line that for almost all \( t,{g}^{\prime }\left( t\right) \) exists and coincides with \( {\chi }_{C}\left( t\right) \) . But the conclusion (11), written in terms of \( g \), says \[ \mathop{\limsup }\limits_{{i \rightarrow \infty }}\frac{g\left( {{y}_{0} + {t}_{i}w}\right) - g\left( {y}_{0}\right) }{{t}_{i}w} < 1. \] Consequently, \( {g}^{\prime }\left( {y}_{0}\right) \) cannot exist and be equal to 1, for any \( {y}_{0} \in C \) . Thus \( C \) has measure 0 . ## Rademacher’s Theorem Theorem 4.16 does not affirm anything in regard to the nonemptiness of \( {\partial }_{D}f\left( {x, y}\right) \) ; we address that issue now. Note to begin with that the gradient of \( f,\nabla f\left( x\right) \), exists for almost all \( x \) when \( f \) is a Lipschitz function on \( {\mathbb{R}}^{n} \) . This is because for each given \( \left( {{x}_{1},{x}_{2},\ldots ,{x}_{i - 1},{x}_{i},{x}_{i + 1},\ldots ,{x}_{n}}\right) \), the function \[ t \mapsto f\left( {{x}_{1},{x}_{2},\ldots ,{x}_{i - 1}, t,{x}_{i + 1},\ldots ,{x}_{n}}\right) \] is Lipschitz on \( \mathbb{R} \), whence \( \partial f/\partial {x}_{i} \) exists for almost all values of \( t \) . Thus the set \( {\Omega }_{i} \) of points \( x \) in \( {\mathbb{R}}^{n} \) at which the partial derivative \( \partial f/\partial {x}_{i} \) fails to exist has the property that the linear measure of \( {\Omega }_{i} \) intersected with any line parallel to the \( i \) th coordinate axis is zero. By Fubini’s Theorem on iterated integration, it follows that \( {\Omega }_{i} \) has measure 0 in \( {\mathbb{R}}^{n} \) . Consequently \( \Omega = \mathop{\bigcup }\limits_{{i = 1}}^{n}{\Omega }_{i} \) has measure 0, and \( \nabla f\left( x\right) \) exists for all \( x \in {\mathbb{R}}^{n} \smallsetminus \Omega \) . The mere existence of \( \nabla f\left( x\right) \) does not imply that it belongs to \( {\partial }_{D}f\left( x\right) \), or even that the latter is nonempty, as seen in Exercise 4.15. But we have the following: 4.18. Corollary. With \( f \) as in the theorem, we have \( \nabla f\left( x\right) \in {\partial }_{D}f\left( x\right) \) for almost all \( x \) in \( {\mathbb{R}}^{n} \) . Proof. The assertion is known in the case \( n = 1 \) (see Exercise 4.15). Let us assume it for dimension \( n \geq 1 \) and derive it for \( n + 1 \) . For all \( x \in {\mathbb{R}}^{n} \) , property (7) holds almost everywhere \( y \) . It follows (from Fubini’s Theorem again) that (7) holds for almost all \( \left( {x, y}\right) \) in \( {\mathbb{R}}^{n} \times \mathbb{R} \) . Similarly, by the induction hypothesis we have, for each \( y \in \mathbb{R} \), for almost all \( x \) , \[ {\nabla }_{x}f\left( {x, y}\right) \in {\partial }_{D}\left( {f\left( {\cdot, y}\right) }\right) \left( x\right) . \] (12) It follows that (12) is valid for almost all \( \left( {x, y}\right) \in {\mathbb{R}}^{n} \times \mathbb{R} \) . Let us now take \( \left( {x, y}\right) \) in that set of full measure in which both (7) and (12) hold. Then, for any \( \left( {v, w}\right) \in {\mathbb{R}}^{n} \times \mathbb{R} \), we have \[ {Df}\left( {x, y;v, w}\right) \geq {Df}\left( {x, y;v,0}\right) + w\frac{\partial f}{\partial y}\left( {x, y}\right) \;\left( {\text{ by }\left( 7\right) }\right) \] \[ \geq \left\langle {{\nabla }_{x}f\left( {x, y}\right), v}\right\rangle + w\frac{\partial f}{\partial y}\left( {x, y}\right) \;\left( {\text{by }\left( {12}\right) }\right) \] \[ = \langle \nabla f\left( {x, y}\right) ,\left( {v, w}\right) \rangle . \] Thus \( \nabla f\left( {x, y}\right) \in {\partial }_{D}f\left( {x, y}\right) \) . The following fact is known as Rademacher's Theorem. 4.19. Corollary. A Lipschitz function on \( {\mathbb{R}}^{n} \) is Fréchet differentiable almost everywhere. Proof. Apply Corollary 4.17 to \( f \) and \( - f \) ; we deduce that for almost all \( x \) , both \( {\partial }_{D}f\left( x\right) \) and \( {\partial }^{D}f\left( x\right) \) are nonempty. But then \( {f}^{\prime }\left( x\right) \) exists for all such \( x \), as noted in Exercise 4.14. We remark that Rademacher's Theorem and the theorem that implied it are essentially local: Their conclusions hold on any open set upon which \( f \) is Lipschitz. This is clear (for example) from the fact that a Lipschitz function on any bounded set can be extended to the whole space so as to be globally Lipschitz (see Problem 1.11.6). ## 5 Sets in \( {L}^{2} \) and Integral Functionals Many of the most important infinite-dimensional applications of nonlinear analysis involve Lebesgue measure and integration. In this section we will study two canonical illustrations of a set and a functional defined in such terms, namely \[ S \mathrel{\text{:=}} \left\{ {x \in {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack : x\left( t\right) \in E\left( t\right) \text{ a.e. }}\right\} \left( {E\left( \cdot \right) \text{ a given multifunction }}\right) ,( \] \[ f\left( x\right) \mathrel{\text{:=}} {\int }_{a}^{b}\varphi \left( {x\left( t\right) }\right) {dt}\left( {x\left( \cdot \right) \in {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack ,\varphi \text{ a given function }}\right) . \] (2) Along the way we will develop the theory of measurable selections of mul-tifunctions, and we will conclude by deriving the nonsmooth analogue of the Euler equation of the classical calculus of variations. Consider first the set \( S \) defined by (1); \( S \) is called a unilateral constraint set in \( {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack = : X \), the underlying Hilbert space throughout this section. 5.1. Exercise. Let \( E\left( t\right) = E\forall t \), where \( E \) is a given closed subset of \( {\mathbb{R}}^{n} \) . Prove that \( S \) is closed, and that \( S \) is convex iff \( E \) is convex. We would like to characterize in terms of the underlying multifunction \( E\left( \cdot \right) \) those vectors \( \zeta \left( \cdot \right) \in {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) which are proximal normals to \( S \) at \( x \in S \) . Accordingly, let \( \zeta \in {N}_{S}^{P}\left( x\right) \) ; then, for some \( \sigma \geq 0 \) we have \[ \left\langle {\zeta ,{x}^{\prime } - x}\right\rangle \leq \sigma {\begin{Vmatrix}{x}^{\prime } - x\end{Vmatrix}}^{2}\forall {x}^{\prime } \in S. \] When the \( {L}^{2} \) inner product and norm are expressed in integral terms, this becomes: \[ {\int }_{a}^{b}\left\{ {\left\langle {-\zeta \left( t\right) ,{x}^{\prime }\left( t\right) - x\left( t\right) }\right\rangle + \sigma {\begin{Vmatrix}{x}^{\prime }\left( t\right) - x\left( t\right) \end{Vmatrix}}^{2}}\right\} {dt} \geq 0\forall {x}^{\prime }\left( \cdot \right) \in S. \] (3) This inequality holds for any measurable and square-integrable function \( {x}^{\prime }\left( \cdot \right) \) taking values in \( E\left( \cdot \right) \) . It is tempting to conclude that (almost everywhere) the integrand is nonnegative on \( E\left( t\right) \) : \[ \left\langle {-\zeta \left( t\right) ,{x}^{\prime } - x\left( t\right) }\right\rangle + \sigma {\begin{Vmatrix}{x}^{\prime } - x\left( t\right) \end{Vmatrix}}^{2} \geq 0\forall {x}^{\prime } \in E\left( t\right) ,\text{ a.e. } \] (4) As we will see, we are justified in yielding to this temptation under mild assumptions. The passage from an integrated to a pointwise conclusion (i.e., from (3) to (4)) can be made rigorous through the theory of measurable selections, which we now pause to develop. We will return later to characterizing \( \zeta \in {N}_{S}^{P}\left( x\right) \), taking up the argument at (3). ## Measurable Multifunctions A multifunction \( \Gamma \) mapping \( {\mathbb{R}}^{m} \) to the subsets of \( {\mathbb{R}}^{n} \) is called measurable provided that the set \[ {\Gamma }^{-1}\left( V\right) \mathrel{\text{:=}} \left\{ {u \in {\mathbb{R}}^{m} : \Gamma \left( u\right) \cap V \neq \varnothing }\right\} \] is (Lebesgue) measurable for every open subset \( V \) of \( {\mathbb{R}}^{n} \) . In order to gain some familiarity with this useful concept, we recommend the following extended exercise. ## 5.2. Exercise. (a) If \( \Gamma \) is measurable, then its domain \( \operatorname{dom}\Gamma \mathrel{\text{:=}} \{ u : \Gamma \left( u\right) \neq \varnothing \} \) is a measurable set, and the multifunction \( \widetilde{\Gamma }\left( u\right) \mathrel{\text{:=}} \operatorname{cl}\Gamma \left( u\right) \) is also measurable. (b) \( \Gamma \) is measurable iff \( {\Gamma }^{-1}\left( V\right) \) is measurable for every closed set \( V \) (or compact set \( V \), or (open or closed) ball \( V \) ). (c) Let \( g : {\mathbb{R}}^{m} \times {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{k} \) be such that for every \( u \in {\mathbb{R}}^{m}, x \mapsto \) \( g\left( {u, x}\right) \) is continuous, and for every \( x \in {\mathbb{R}}^{n}, u \mapsto g\left( {u, x}\right) \) is measurable. Set \[ \Gamma \left( u\right) \mathrel{
1012_(GTM178)Nonsmooth Analysis and Control Theory
53
r every open subset \( V \) of \( {\mathbb{R}}^{n} \) . In order to gain some familiarity with this useful concept, we recommend the following extended exercise. ## 5.2. Exercise. (a) If \( \Gamma \) is measurable, then its domain \( \operatorname{dom}\Gamma \mathrel{\text{:=}} \{ u : \Gamma \left( u\right) \neq \varnothing \} \) is a measurable set, and the multifunction \( \widetilde{\Gamma }\left( u\right) \mathrel{\text{:=}} \operatorname{cl}\Gamma \left( u\right) \) is also measurable. (b) \( \Gamma \) is measurable iff \( {\Gamma }^{-1}\left( V\right) \) is measurable for every closed set \( V \) (or compact set \( V \), or (open or closed) ball \( V \) ). (c) Let \( g : {\mathbb{R}}^{m} \times {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{k} \) be such that for every \( u \in {\mathbb{R}}^{m}, x \mapsto \) \( g\left( {u, x}\right) \) is continuous, and for every \( x \in {\mathbb{R}}^{n}, u \mapsto g\left( {u, x}\right) \) is measurable. Set \[ \Gamma \left( u\right) \mathrel{\text{:=}} \left\{ {x \in {\mathbb{R}}^{n} : g\left( {u, x}\right) = 0}\right\} . \] Prove that \( \Gamma \) is measurable. (Hint. Let \( V \) be compact, and let \( \left\{ {v}_{i}\right\} \) be a countable dense set in \( V \) ; show that we have \[ \left. {{\Gamma }^{-1}\left( V\right) = \mathop{\bigcap }\limits_{{i = 1}}^{\infty }\mathop{\bigcup }\limits_{{j = 1}}^{\infty }\left\{ {u : \left| {g\left( {u,{v}_{j}}\right) }\right| < \frac{1}{i}}\right\} .}\right) \] (d) \( \Gamma \) is called closed-valued if \( \Gamma \left( u\right) \) is a closed set for every \( u \in {\mathbb{R}}^{m} \) . Prove that when \( \Gamma \) is closed-valued, then \( \Gamma \) is measurable iff the function \( u \mapsto d\left( {x,\Gamma \left( u\right) }\right) \) mapping \( {\mathbb{R}}^{m} \) to \( \left\lbrack {0,\infty }\right\rbrack \) is measurable for each \( x \in {\mathbb{R}}^{n} \) . (e) The graph of \( \Gamma ,\operatorname{gr}\Gamma \), is of course the set \[ \left\{ {\left( {u, x}\right) \in {\mathbb{R}}^{m} \times {\mathbb{R}}^{n} : x \in \Gamma \left( u\right) }\right\} . \] Prove that \( \Gamma \) is measurable if \( \operatorname{gr}\Gamma \) is closed. (f) A function \( \gamma : {\mathbb{R}}^{m} \rightarrow {\mathbb{R}}^{n} \) is measurable iff the multifunction \( \Gamma \left( u\right) \mathrel{\text{:=}} \{ \gamma \left( u\right) \} \) is measurable, and when \( n = 1 \), iff \( \Gamma \left( u\right) \mathrel{\text{:=}} \) \( \lbrack \gamma \left( t\right) ,\infty ) \) is measurable. (g) Let \( \Gamma \) have closed graph, and let \( \theta : {\mathbb{R}}^{\ell } \rightarrow {\mathbb{R}}^{m} \) be measurable. Prove that the multifunction \( w \mapsto \Gamma \left( {\theta \left( w\right) }\right) \) is measurable. (Hint. \[ \left. {\left\{ {w : \Gamma \left( {\theta \left( w\right) }\right) \cap V \neq \varnothing }\right\} = {\theta }^{-1}\left( {{\Gamma }^{-1}\left( V\right) }\right) .}\right) \] Use (f) to deduce that \( \gamma \left( {\theta \left( \cdot \right) }\right) \) is measurable if \( \gamma : {\mathbb{R}}^{m} \rightarrow \mathbb{R} \) is lower semicontinuous. (h) If \( {\Gamma }_{1} \) and \( {\Gamma }_{2} \) are two closed-valued measurable multifunctions, then \( \Gamma \left( u\right) \mathrel{\text{:=}} {\Gamma }_{1}\left( u\right) \cap {\Gamma }_{2}\left( u\right) \) defines another closed-valued measurable multifunction. (Hint. Let \( V,\left\{ {v}_{i}\right\} \) be as in (c), and observe \[ \{ u : \Gamma \left( u\right) \cap V \neq \varnothing \} \] \[ \left. { = \mathop{\bigcap }\limits_{{i = 1}}^{\infty }\mathop{\bigcup }\limits_{{j = 1}}^{\infty }\left\{ {u : d\left( {{v}_{j},{\Gamma }_{1}\left( u\right) }\right) + d\left( {{v}_{j},{\Gamma }_{2}\left( u\right) }\right) < \frac{1}{i}}\right\} .}\right) \] (i) Let \( g \) be as in (c), with \( k = 1 \) . Prove that for any scalars \( c \) and \( d \), the following multifunction is measurable: \[ \Gamma \left( u\right) \mathrel{\text{:=}} \left\{ {x \in {\mathbb{R}}^{n} : c \leq g\left( {u, x}\right) \leq d}\right\} . \] 5.3. Theorem (Measurable Selection). Let \( \Gamma \) be closed-valued and measurable. Then there exists a measurable function \( \gamma \) such that \[ \gamma \left( u\right) \in \Gamma \left( u\right) \forall u \in \operatorname{dom}\Gamma . \] Proof. Let \( \Delta \mathrel{\text{:=}} \operatorname{dom}\Gamma \) . We begin by noting that for any \( \zeta \) in \( {\mathbb{R}}^{n} \) the function \( s \rightarrow {d}_{\Gamma \left( s\right) }\left( \zeta \right) \) is measurable on \( \Delta \) (where \( {d}_{\Gamma \left( s\right) } \) is as usual the Euclidean distance function), since \[ \left\{ {s \in \Delta : {d}_{\Gamma \left( s\right) }\left( \zeta \right) \leq \alpha }\right\} = \{ s \in \Delta : \Gamma \left( s\right) \cap \left\lbrack {\zeta + \alpha \bar{B}}\right\rbrack \neq \varnothing \} . \] Now let \( \left\{ {\zeta }_{i}\right\} \) be a countable dense subset of \( {\mathbb{R}}^{n} \), and define a function \( {\gamma }_{0} : \Delta \rightarrow {R}^{n} \) as follows: \[ {\gamma }_{0}\left( s\right) = \text{ the first }{\zeta }_{i}\text{ such that }{d}_{\Gamma \left( s\right) }\left( {\zeta }_{i}\right) \leq 1. \] Lemma. The functions \( s \rightarrow {\gamma }_{0}\left( s\right) \) and \( s \rightarrow {d}_{\Gamma \left( s\right) }\left( {{\gamma }_{0}\left( s\right) }\right) \) are measurable. To see this, observe that \( {\gamma }_{0} \) assumes countably many values, and that, for each \( i \) , \[ \left\{ {s : {\gamma }_{0}\left( s\right) = {\zeta }_{i}}\right\} = \mathop{\bigcap }\limits_{j}\left\{ {s : {d}_{\Gamma \left( s\right) }\left( {\zeta }_{j}\right) > 1}\right\} \cap \left\{ {s : {d}_{\Gamma \left( s\right) }\left( {\zeta }_{i}\right) \leq 1}\right\} , \] where the intersection is over \( j = 1,\ldots, i - 1 \) . This implies that \( {\gamma }_{0} \) is measurable. To complete the proof of the lemma, we need only note \[ \left\{ {s : {d}_{\Gamma \left( s\right) }\left( {{\gamma }_{0}\left( s\right) }\right) > \alpha }\right\} = \mathop{\bigcup }\limits_{j}\left\lbrack {\left\{ {s : {\gamma }_{0}\left( s\right) = {\zeta }_{j}}\right\} \cap \left\{ {s : {d}_{\Gamma \left( s\right) }\left( {\zeta }_{j}\right) > \alpha }\right\} }\right\rbrack , \] where the union is over the positive integers \( j \) . We pursue the process begun above by defining for each integer \( i \) a function \( {\gamma }_{i + 1} \) such that \( {\gamma }_{i + 1}\left( s\right) \) is the first \( {\zeta }_{j} \) for which both the following hold: \[ \begin{Vmatrix}{{\zeta }_{j} - {\gamma }_{i}\left( s\right) }\end{Vmatrix} \leq \frac{2}{3}{d}_{\Gamma \left( s\right) }\left( {{\gamma }_{i}\left( s\right) }\right) ,\;{d}_{\Gamma \left( s\right) }\left( {\zeta }_{j}\right) \leq \frac{2}{3}{d}_{\Gamma \left( s\right) }\left( {{\gamma }_{i}\left( s\right) }\right) . \] It follows much as above that each \( {\gamma }_{i} \) is measurable. Furthermore, we deduce the inequalities \[ {d}_{\Gamma \left( s\right) }\left( {{\gamma }_{i + 1}\left( s\right) }\right) \leq {\left( \frac{2}{3}\right) }^{i}{d}_{\Gamma \left( s\right) }\left( {{\gamma }_{0}\left( s\right) }\right) \leq {\left( \frac{2}{3}\right) }^{i}, \] together with \( \begin{Vmatrix}{{\gamma }_{i + 1}\left( s\right) - {\gamma }_{i}\left( s\right) }\end{Vmatrix} \leq {\frac{2}{3}}^{i + 1} \) . It follows that \( \left\{ {{\gamma }_{i}\left( s\right) }\right\} \) is a Cauchy sequence converging to a value \( \gamma \left( s\right) \) for each \( s \), and that \( \gamma \) is a measurable selection for \( \Gamma \) . ## 5.4. Exercise. (a) Let \( \varphi : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be a Lipschitz function, and let \( x\left( t\right) \) be a measurable mapping from \( \mathbb{R} \) to \( {\mathbb{R}}^{n} \) . Prove that the multifunctions \( t \mapsto {\partial }_{L}\varphi \left( {x\left( t\right) }\right) \) and \( t \mapsto {\partial }_{C}\varphi \left( {x\left( t\right) }\right) \) are measurable. (b) Let the multifunction \( \Gamma \) from \( \mathbb{R} \) to \( {\mathbb{R}}^{n} \) be compact-valued and measurable, and let \( v : \mathbb{R} \rightarrow {\mathbb{R}}^{n} \) be measurable. Prove the existence of a measurable selection \( \gamma \) for \( \Gamma \) such that, for all \( t \in \operatorname{dom}\Gamma \), we have \[ \langle \gamma \left( t\right), v\left( t\right) \rangle = \max \left\{ {\left\langle {{\gamma }^{\prime }, v\left( t\right) }\right\rangle : {\gamma }^{\prime } \in \Gamma \left( t\right) }\right\} . \] In the following, the multifunction \( E\left( \cdot \right) \) and the subset \( S \) of \( {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) are related as in (1). 5.5. Corollary. Let \( g : \left\lbrack {a, b}\right\rbrack \times {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be such that \( t \mapsto g\left( {t, x}\right) \) is measurable for each \( x \), and \( x \mapsto g\left( {t, x}\right) \) is continuous for \( t \) a.e. Let the multifunction \( E\left( \cdot \right) \) be measurable and closed-valued. Suppose that we have \[ {\int }_{a}^{b}g\left( {t, x\left( t\right) }\right) {dt} \geq 0 \] whenever \( x\left( \cdot \right) \) belongs to \( S \) and the integral is defined, and suppose that for a certain \( {x}_{0}\left( \cdot \right) \in S \) we have \( g\left( {t,{x}_{0}\left( t\right) }\right) = 0 \) a.e. Then \[ g\left( {t, x}\right) \geq 0\forall x \in E\left( t\right), t\text{ a.e. } \] Proof. It suffices to prove that for any \( k > 0 \), the following set has measure 0: \[ \{ t : \text{ for some }x \in E\left( t\right) \cap \bar{B}\left( {0;k}\right) \text{ we have } - k \leq g\left( {t, x}\right) \leq - 1/k\} \] (why?). If this is not the case, then the domain of the following multifunction \( \Gamma \) is of positive measure: \[ \Gamma \left( t\right) \mathrel{\text{:=}} \{ x \in E\left( t\right) \cap \bar{B}\left( {0, k}\right) : - k \leq g\left( {t, x}\right) \leq - 1/k\} . \
1012_(GTM178)Nonsmooth Analysis and Control Theory
54
a.e. Let the multifunction \( E\left( \cdot \right) \) be measurable and closed-valued. Suppose that we have \[ {\int }_{a}^{b}g\left( {t, x\left( t\right) }\right) {dt} \geq 0 \] whenever \( x\left( \cdot \right) \) belongs to \( S \) and the integral is defined, and suppose that for a certain \( {x}_{0}\left( \cdot \right) \in S \) we have \( g\left( {t,{x}_{0}\left( t\right) }\right) = 0 \) a.e. Then \[ g\left( {t, x}\right) \geq 0\forall x \in E\left( t\right), t\text{ a.e. } \] Proof. It suffices to prove that for any \( k > 0 \), the following set has measure 0: \[ \{ t : \text{ for some }x \in E\left( t\right) \cap \bar{B}\left( {0;k}\right) \text{ we have } - k \leq g\left( {t, x}\right) \leq - 1/k\} \] (why?). If this is not the case, then the domain of the following multifunction \( \Gamma \) is of positive measure: \[ \Gamma \left( t\right) \mathrel{\text{:=}} \{ x \in E\left( t\right) \cap \bar{B}\left( {0, k}\right) : - k \leq g\left( {t, x}\right) \leq - 1/k\} . \] Further, \( \Gamma \) is measurable, in view of Exercise 5.2(h, i). We invoke Theorem 5.3, the Measurable Selection Theorem, to deduce the existence of a selection \( x\left( \cdot \right) \) for \( \Gamma \) on dom \( \Gamma \) . We extend \( x\left( \cdot \right) \) to all of \( \left\lbrack {a, b}\right\rbrack \) by defining \( x\left( t\right) = {x}_{0}\left( t\right) \) for \( t \in \left\lbrack {a, b}\right\rbrack \smallsetminus \operatorname{dom}\Gamma \) . Then \( x\left( \cdot \right) \) belongs to \( S \), the function \( x \mapsto g\left( {t, x\left( t\right) }\right) \) is integrable, and its integral over \( \left\lbrack {a, b}\right\rbrack \) is strictly negative. This contradiction proves the corollary. Returning now to the inequality (3), and assuming henceforth that \( E\left( \cdot \right) \) is measurable and closed-valued, we are ready to pursue the analysis of \( {N}_{S}^{P}\left( x\right) \) . 5.6. Exercise. Show that if (3) holds, then (4) does, so that \[ \zeta \left( t\right) \in {N}_{E\left( t\right) }^{P}\left( {x\left( t\right) }\right) \text{ a.e. } \] The condition (4) that we have obtained differs from the pointwise inclusion \( \zeta \left( t\right) \in {N}_{E\left( t\right) }^{P}\left( {x\left( t\right) }\right) \) in one important respect: The latter implies that for almost all \( t \), there exists \( \sigma \left( t\right) \geq 0 \) such that \[ \left\langle {-\zeta \left( t\right) ,{x}^{\prime } - x\left( t\right) }\right\rangle + \sigma \left( t\right) {\begin{Vmatrix}{x}^{\prime } - x\left( t\right) \end{Vmatrix}}^{2} \geq 0\forall {x}^{\prime } \in E\left( t\right) . \] But in (4) the same \( \sigma \) independent of \( t \) appears. To take account of this, let us denote by \( {N}_{E}^{P,\sigma }\left( x\right) \) those vectors \( \zeta \) satisfying \[ \left\langle {\zeta ,{x}^{\prime } - x}\right\rangle \leq \sigma {\begin{Vmatrix}{x}^{\prime } - x\end{Vmatrix}}^{2}\forall {x}^{\prime } \in E. \] 5.7. Proposition. \( \zeta \in {N}_{S}^{P}\left( x\right) \) iff for some \( \sigma \geq 0 \), we have \[ \zeta \left( t\right) \in {N}_{E\left( t\right) }^{P,\sigma }\left( {x\left( t\right) }\right) \text{ a.e. } \] 5.8. Exercise. Prove Proposition 5.7. ## An Integral Functional We turn now to the integral functional \( f \) on \( X \mathrel{\text{:=}} {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) defined by \[ f\left( x\right) \mathrel{\text{:=}} {\int }_{a}^{b}\varphi \left( {x\left( t\right) }\right) {dt} \] where for simplicity we will suppose that \( \varphi \) is a globally Lipschitz function on \( {\mathbb{R}}^{n} \), with Lipschitz constant \( K \) . 5.9. Exercise. Prove that \( f \) is well defined and finite on \( X \), and globally Lipschitz with Lipschitz constant \( K{\left( b - a\right) }^{1/2} \) . Our first interest is to study \( {\partial }_{P}f \) ; accordingly, let \( \zeta \in {\partial }_{P}f\left( x\right) \), and let us write the corresponding proximal subgradient inequality: For some \( \sigma \geq 0 \) and \( \eta > 0 \) we have \[ {\int }_{a}^{b}\left\{ {\varphi \left( {y\left( t\right) }\right) - \varphi \left( {x\left( t\right) }\right) + \sigma \parallel y\left( t\right) - x\left( t\right) {\parallel }^{2}-\langle \zeta \left( t\right), y\left( t\right) - x\left( t\right) \rangle }\right\} {dt} \geq 0 \] whenever \( \parallel y - x{\parallel }_{2} < \eta \) . What can be deduced from this? 5.10. Theorem. If \( \zeta \in {\partial }_{P}f\left( x\right) \), then \( \zeta \left( t\right) \in {\partial }_{P}\varphi \left( {x\left( t\right) }\right) \) a.e. We remark that the theorem implies that for any \( v\left( \cdot \right) \in X \), we have \[ \langle \zeta, v\rangle = {\int }_{a}^{b}\langle \zeta \left( t\right), v\left( t\right) \rangle {dt} \] where \( \zeta \left( t\right) \in {\partial }_{P}\varphi \left( {x\left( t\right) }\right) \) a.e. In abridged notation, this might be written \( {\partial }_{P}\int \varphi = \int {\partial }_{P}\varphi \), which reveals the proposition as an analogue of the technique known as "differentiating under the integral." Proof. Let any \( M > 0 \) be given. We will prove that for almost all \( t \), we have \[ \varphi \left( y\right) - \varphi \left( {x\left( t\right) }\right) + \sigma \parallel y - x\left( t\right) {\parallel }^{2} - \langle \zeta \left( t\right), y - x\left( t\right) \rangle \geq 0\forall y \in M\bar{B}, \] (6) which, since \( M \) is arbitrary, leads to the desired conclusion. Pick \( c, d \) in \( \left( {a, b}\right) \) with \( c < d \) such that \[ M{\left( d - c\right) }^{1/2} + {\left\{ {\int }_{c}^{d}\parallel x\left( t\right) {\parallel }^{2}dt\right\} }^{1/2} < \eta \] where \( \eta \) is defined in connection with inequality (5). We define \[ \widehat{\varphi }\left( {t, y}\right) \mathrel{\text{:=}} \left\{ \begin{array}{ll} 0 & \text{ if }t \notin \left\lbrack {c, d}\right\rbrack , \\ \varphi \left( y\right) - \varphi \left( {x\left( t\right) }\right) + \sigma \parallel y - x\left( t\right) {\parallel }^{2} - \langle \zeta \left( t\right), y - x\left( t\right) \rangle & \text{ if }t \in \left\lbrack {c, d}\right\rbrack . \end{array}\right. \] Let \( E \mathrel{\text{:=}} M\bar{B} \), and let \( y\left( \cdot \right) \) be any element of \( {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) satisfying \( y\left( t\right) \in E \) a.e. Then, if \( \widehat{y} \) is the element of \( X \) defined by \[ \widehat{y}\left( t\right) \mathrel{\text{:=}} \left\{ \begin{array}{ll} x\left( t\right) & \text{ if }t \notin \left\lbrack {c, d}\right\rbrack \\ y\left( t\right) & \text{ if }t \in \left\lbrack {c, d}\right\rbrack \end{array}\right. \] we have \[ \parallel \widehat{y} - x\parallel = {\left\{ {\int }_{c}^{d}\parallel y\left( t\right) - x\left( t\right) {\parallel }^{2}dt\right\} }^{1/2} \] \[ \leq {\left\{ {\int }_{c}^{d}{\begin{Vmatrix}y\left( t\right) \end{Vmatrix}}^{2}dt\right\} }^{1/2} + {\left\{ {\int }_{c}^{d}{\begin{Vmatrix}x\left( t\right) \end{Vmatrix}}^{2}dt\right\} }^{1/2} \] \[ \leq M{\left( d - c\right) }^{1/2} + {\left\{ {\int }_{c}^{d}\parallel x\left( t\right) {\parallel }^{2}dt\right\} }^{1/2} < \eta . \] Thus (5) may be invoked to yield \[ {\int }_{a}^{b}\widehat{\varphi }\left( {t, y\left( t\right) }\right) {dt} = {\int }_{c}^{d}\widehat{\varphi }\left( {t, y\left( t\right) }\right) {dt} \] \[ = {\int }_{c}^{d}\left\{ {\varphi \left( {y\left( t\right) }\right) - \varphi \left( {x\left( t\right) }\right) + \sigma \parallel y\left( t\right) - x\left( t\right) {\parallel }^{2}-\langle \zeta \left( t\right), y\left( t\right) - x\left( t\right) \rangle }\right\} {dt} \] \[ = {\int }_{a}^{b}\left\{ {\varphi \left( {\widehat{y}\left( t\right) }\right) - \varphi \left( {x\left( t\right) }\right) + \sigma \parallel \widehat{y}\left( t\right) - x\left( t\right) {\parallel }^{2}-\langle \zeta \left( t\right) ,\widehat{y}\left( t\right) - x\left( t\right) \rangle }\right\} {dt} \] \[ \geq 0\text{.} \] Note also that \( \widehat{\varphi }\left( {t, x\left( t\right) }\right) = 0 \) a.e. Since \( \widehat{\varphi } \) is measurable in \( t \) and continuous in \( y \), it follows now from Corollary 5.5 that for \( t \) a.e., we have \[ \widehat{\varphi }\left( {t, y}\right) \geq 0\forall y \in E = M\bar{B} \] Consequently,(6) holds for \( t \in \left\lbrack {c, d}\right\rbrack \) a.e. But since \( \left\lbrack {c, d}\right\rbrack \) is an arbitrary (small) subinterval of \( \left\lbrack {a, b}\right\rbrack \) ,(6) in fact holds a.e. in \( \left\lbrack {a, b}\right\rbrack \), as claimed. ## 5.11. Exercise. (a) Let \( {\partial }_{P}^{\sigma }\varphi \left( x\right) \) denote the "global proximal subgradients of rank \( \sigma \) "; i.e., those \( \zeta \) satisfying \[ \varphi \left( y\right) - \varphi \left( x\right) + \sigma \parallel y - x{\parallel }^{2} \geq \langle \zeta, y - x\rangle \forall y \in {\mathbb{R}}^{n}. \] Show that in Theorem 5.10, the conclusion may be strengthened to \( \zeta \left( t\right) \in {\partial }_{P}^{\sigma }\varphi \left( {x\left( t\right) }\right) \) a.e., and that conversely this last property (for a given \( \sigma \geq 0 \) ) implies that \( \zeta \) belongs to \( {\partial }_{P}f\left( x\right) \) . (b) If \( \zeta \in {\partial }_{P}\varphi \left( x\right) \), then show that for some \( \sigma > 0 \) we actually have \( \zeta \in {\partial }_{P}^{\sigma }\varphi \left( x\right) \) . (This uses the global Lipschitz hypothesis on \( \varphi \) .) The exercise shows that a certain global and uniform character of the pointwise condition \( \zeta \left( t\right) \in {\partial }_{P}\varphi \left( {x\left( t\right) }\right) \) is required to imply \( \zeta \in {\partial }_{P}f\left( x\right) \) . Nonetheless, an approximate converse to Theorem 5.10 is provided by the following. 5.12. Proposition. Let \( \zeta \in {L}_{n}^{2}\left\lbrack {0,1}\right\rbrack \) be such that \( \zeta \left( t\right) \in {\partial
1012_(GTM178)Nonsmooth Analysis and Control Theory
55
may be strengthened to \( \zeta \left( t\right) \in {\partial }_{P}^{\sigma }\varphi \left( {x\left( t\right) }\right) \) a.e., and that conversely this last property (for a given \( \sigma \geq 0 \) ) implies that \( \zeta \) belongs to \( {\partial }_{P}f\left( x\right) \) . (b) If \( \zeta \in {\partial }_{P}\varphi \left( x\right) \), then show that for some \( \sigma > 0 \) we actually have \( \zeta \in {\partial }_{P}^{\sigma }\varphi \left( x\right) \) . (This uses the global Lipschitz hypothesis on \( \varphi \) .) The exercise shows that a certain global and uniform character of the pointwise condition \( \zeta \left( t\right) \in {\partial }_{P}\varphi \left( {x\left( t\right) }\right) \) is required to imply \( \zeta \in {\partial }_{P}f\left( x\right) \) . Nonetheless, an approximate converse to Theorem 5.10 is provided by the following. 5.12. Proposition. Let \( \zeta \in {L}_{n}^{2}\left\lbrack {0,1}\right\rbrack \) be such that \( \zeta \left( t\right) \in {\partial }_{P}\varphi \left( {x\left( t\right) }\right) \) a.e. Then for any \( \varepsilon > 0 \) there exist \( {x}^{\prime } \in X \) and \( {\zeta }^{\prime } \in {\partial }_{P}f\left( {x}^{\prime }\right) \) with \( \begin{Vmatrix}{{\zeta }^{\prime } - \zeta }\end{Vmatrix} < \varepsilon \) , \( \begin{Vmatrix}{{x}^{\prime } - x}\end{Vmatrix} < \varepsilon \) Proof. According to Exercise 5.11(b), there exists for almost each \( t \) a number \( \sigma \left( t\right) \) such that \[ \zeta \left( t\right) \in {\partial }_{P}^{\sigma \left( t\right) }\varphi \left( {x\left( t\right) }\right) \text{ a.e. } \] In fact, let us take \( \sigma \left( t\right) \) equal to the first positive integer \( k \) such that \( \zeta \left( t\right) \in {\partial }_{P}^{k}\varphi \left( {x\left( t\right) }\right) \) . Then, for any positive integer \( k \), we have \[ \{ t \in \left\lbrack {a, b}\right\rbrack : \sigma \left( t\right) = k\} \] \[ = \left\{ {t : \varphi \left( y\right) - \varphi \left( {x\left( t\right) }\right) + k\parallel y - x\left( t\right) {\parallel }^{2} \geq \left\langle {\zeta \left( t\right), y - x\left( t\right) }\right\rangle \;\forall y \in {\mathbb{R}}^{n}}\right\} \] \[ \mathop{\bigcap }\limits_{{j = 1}}^{{k - 1}}\left\{ {t : \varphi \left( y\right) - \varphi \left( {x\left( t\right) }\right) + j\parallel y - x\left( t\right) {\parallel }^{2} < \langle \zeta \left( t\right), y - x\left( t\right) \rangle \text{ for some }y \in {\mathbb{R}}^{n}}\right\} . \] This implies that the function \( \sigma \left( \cdot \right) \) is measurable. Now let \( \left( {\bar{x},\bar{\zeta },\bar{\sigma }}\right) \) be any triple such that \( \bar{\zeta } \in {\partial }_{P}^{\bar{\sigma }}\varphi \left( \bar{x}\right) \) (Why is there such a triple?). For every \( M > \bar{\sigma } \) , let \( {\Omega }_{M} \) be the set \( \{ t : \sigma \left( t\right) < M\} \), and define \[ {x}^{\prime }\left( t\right) = \left\{ {\begin{array}{ll} x\left( t\right) & \text{ if }t \in {\Omega }_{M}, \\ \bar{x} & \text{ otherwise,} \end{array}\;{\zeta }^{\prime }\left( t\right) = \left\{ \begin{array}{ll} \zeta \left( t\right) & \text{ if }t \in {\Omega }_{M}, \\ \bar{\zeta } & \text{ otherwise. } \end{array}\right. }\right. \] Then \( {\zeta }^{\prime }\left( t\right) \in {\partial }_{P}^{M}\varphi \left( {{x}^{\prime }\left( t\right) }\right) \) for \( t \) a.e., whence \( {\zeta }^{\prime } \in {\partial }_{P}f\left( {x}^{\prime }\right) \) by Exercise 5.11(a). Since meas \( \left( {\Omega }_{M}\right) \rightarrow \left( {b - a}\right) \) as \( M \rightarrow \infty \), it follows that \( {x}^{\prime } \rightarrow x \) and \( {\zeta }^{\prime } \rightarrow \zeta \) in \( X \) as \( M \rightarrow \infty \), a final step in the proof that we ask the reader to verify. ## A Convexity Property As witnessed above, the correspondence between the proximal subgradients of the integral functional \( f \) and those of the generating integrand \( \varphi \) is quite a close one. We turn now to consideration of the limiting subdifferential \( {\partial }_{L}f\left( x\right) \) . We will discover in Theorem 5.18 below an interesting convexifica-tion property of the weak-closure operation that generates \( {\partial }_{L}f \) from \( {\partial }_{P}f \) . To begin understanding this property, consider the natural conjecture that if \( \zeta \in {\partial }_{L}f\left( x\right) \), then \( \zeta \left( t\right) \in {\partial }_{L}\varphi \left( {x\left( t\right) }\right) \) a.e. (We call this "natural" in view of the fact that we have established this fact when \( {\partial }_{L} \) is replaced by \( {\partial }_{P} \) .) It turns out that this conjecture is wrong, and the following example will lend insight as to why. 5.13. Exercise. We set \( \varphi \left( x\right) = - \left| x\right| \), with \( n = 1 \) . We will construct a sequence \( {x}_{i} \) in \( X \) converging to 0, with elements \( {\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) \) such that \( \left\{ {\zeta }_{i}\right\} \) converges weakly to 0 . It follows that \( 0 \in {\partial }_{L}f\left( 0\right) \), yet it is not the case that 0 belongs to \( {\partial }_{L}\varphi \left( 0\right) = \{ - 1,1\} \) . We define a function \( {x}_{i}\left( t\right) \) on \( \left\lbrack {0,1}\right\rbrack \) by setting \( {x}_{i}\left( t\right) = 1/i \) on each subinterval of the form \( \left( {k/i,\left( {k + 1}\right) /i}\right) \), where \( k \) is an even integer in \( \left\lbrack {0, i - 1}\right\rbrack \), and \( {x}_{i}\left( t\right) = - 1/i \) elsewhere. Then \( {x}_{i} \rightarrow 0 \) in \( X \) . Let \( {\zeta }_{i}\left( t\right) = - {x}_{i}\left( t\right) /\left| {{x}_{i}\left( t\right) }\right| \) . (a) Prove that \( {\zeta }_{i} \) converges weakly to 0 in \( X \) . (b) Prove that \( {\zeta }_{i} \in {\partial }_{P}f\left( {x}_{i}\right) \) (use Exercise 5.11). In the foregoing example, observe that although \( 0 \notin {\partial }_{L}\varphi \left( 0\right) = \{ - 1,1\} \), we do have \( 0 \in \operatorname{co}{\partial }_{L}\varphi \left( 0\right) = {\partial }_{C}\varphi \left( 0\right) \) . We will obtain a general conclusion along these lines below. The key to the analysis is a certain convexification property of integration relative to (nonatomic) measures, a phenomenon first observed by A. M. Lyapounov but reflected especially clearly below in a more recent theorem due to Aumann. A multifunction \( F \) mapping \( \left\lbrack {a, b}\right\rbrack \) to the subsets of \( {\mathbb{R}}^{n} \) is said to be \( {L}^{2} \) - bounded if there is a function \( k \in {L}_{1}^{2}\left\lbrack {a, b}\right\rbrack \) such that \[ \parallel v\parallel \leq k\left( t\right) \forall v \in F\left( t\right) \text{, a.e. } \] By \( \operatorname{co}F \) we mean the multifunction whose value at \( t \) is \( \operatorname{co}F\left( t\right) \) . Finally, \( \int F \) signifies the set of all points of the form \( {\int }_{a}^{b}f\left( t\right) {dt} \), where \( f\left( \cdot \right) \) is an integrable selection of \( F \) . 5.14. Exercise. Let the multifunction \( E\left( \cdot \right) \) of (1) be \( {L}^{2} \) -bounded, and have convex closed values. Prove that the set \( S \) of its measurable selections is weakly compact, and that \( \int E \) is a compact convex subset of \( {\mathbb{R}}^{n} \) . 5.15. Theorem (Aumann). Let \( F \) be a measurable, \( {L}^{2} \) -bounded multifunction from \( \left\lbrack {a, b}\right\rbrack \) to \( {\mathbb{R}}^{n} \) whose values are closed and nonempty. Then \[ \int F = \int \operatorname{co}F \] Step 1. It suffices to show that any point \( \xi \) lying in \( \int \operatorname{co}F \) also lies in \( \int F \) ; we may take \( \xi = 0 \) without loss of generality. Exercise 5.14 shows that \( \int \operatorname{co}F \) is convex compact. Its dimension \( K \) is defined to be the dimension of the minimal subspace \( L \) of \( {\mathbb{R}}^{n} \) containing it. If \( K = 0,\int \operatorname{co}F \) is a point and must coincide with \( \int F \), since the latter is nonempty as a consequence of the Measurable Selection Theorem 5.3. So we may assume \( K \geq 1 \), and that the theorem is true for multifunctions \( G \) such that \( \dim \int \operatorname{co}G \leq K - 1 \) . The proof will be by induction. Step 2. Suppose now that 0 does not lie in the relative interior of \( \int \operatorname{co}F \) (i.e., interior relative to \( L \) ). Then there is a nonzero vector \( d \) in \( L \) normal to \( \int \operatorname{co}F \) at 0 ; that is, such that \[ \langle d, w\rangle \leq 0\forall w \in \int \operatorname{co}F. \] Let \( S \) be the set of measurable (and necessarily, square integrable) selections of \( \operatorname{co}F \) on \( \left\lbrack {a, b}\right\rbrack \), and let \( {s}_{0} \) be an element of \( S \) satisfying \( {\int }_{a}^{b}{s}_{0}\left( t\right) {dt} = 0 \) . Then \[ {\int }_{a}^{b}\langle d, s\left( t\right) \rangle {dt} \leq 0 = {\int }_{a}^{b}\left\langle {d,{s}_{0}\left( t\right) }\right\rangle {dt}\forall s \in S. \] Let \( H\left( {\cdot ;F\left( t\right) }\right) \) designate the upper support function of \( F\left( t\right) \), or equivalently, of \( \operatorname{co}F\left( t\right) \) : \[ H\left( {x;F\left( t\right) }\right) \mathrel{\text{:=}} \max \{ \langle x, f\rangle : f \in F\left( t\right) \} . \] It follows from Exercise 5.4(b) that we have \[ \mathop{\max }\limits_{{s \in S}}{\int }_{a}^{b}\langle d, s\left( t\right) \rangle {dt} = {\int }_{a}^{b}H\left( {d;F\left( t\right) }\right) {dt}. \] We deduce \[ {\int }_{a}^{b}\left\langle {d,{s}_{0}\left( t\right) }\right\rangle \leq {\int }_{b}^{b}H\left( {d;F\left( t\right) }\right) {dt} \leq 0 = {\int }_{a}^{b}\left\langle {d,{s}_{0}\left( t\right) }\right\rangle {dt} \] whence \[ H\left( {d;F\left( t\right) }\right) = \left\langle {d,{s}_{0}\left( t\right) }\right\rangle \text{ a.e. } \] Let us define a new multifunction
1012_(GTM178)Nonsmooth Analysis and Control Theory
56
{dt} \leq 0 = {\int }_{a}^{b}\left\langle {d,{s}_{0}\left( t\right) }\right\rangle {dt}\forall s \in S. \] Let \( H\left( {\cdot ;F\left( t\right) }\right) \) designate the upper support function of \( F\left( t\right) \), or equivalently, of \( \operatorname{co}F\left( t\right) \) : \[ H\left( {x;F\left( t\right) }\right) \mathrel{\text{:=}} \max \{ \langle x, f\rangle : f \in F\left( t\right) \} . \] It follows from Exercise 5.4(b) that we have \[ \mathop{\max }\limits_{{s \in S}}{\int }_{a}^{b}\langle d, s\left( t\right) \rangle {dt} = {\int }_{a}^{b}H\left( {d;F\left( t\right) }\right) {dt}. \] We deduce \[ {\int }_{a}^{b}\left\langle {d,{s}_{0}\left( t\right) }\right\rangle \leq {\int }_{b}^{b}H\left( {d;F\left( t\right) }\right) {dt} \leq 0 = {\int }_{a}^{b}\left\langle {d,{s}_{0}\left( t\right) }\right\rangle {dt} \] whence \[ H\left( {d;F\left( t\right) }\right) = \left\langle {d,{s}_{0}\left( t\right) }\right\rangle \text{ a.e. } \] Let us define a new multifunction \( \widetilde{F} \) via \[ \widetilde{F}\left( t\right) \mathrel{\text{:=}} \{ f \in F\left( t\right) : \langle f, d\rangle = H\left( {d;F\left( t\right) }\right) \} . \] 5.16. Exercise. Show that \( \widetilde{F} \) is nonempty and closed-valued, measurable on \( \left\lbrack {a, b}\right\rbrack \), and that \( {s}_{0}\left( t\right) \in \operatorname{co}\widetilde{F}\left( t\right) \) a.e. It follows that \( \widetilde{F} \) satisfies all the hypotheses of the theorem, and that \( 0 \in \) \( \int \operatorname{co}\widetilde{F} \) . Further \( \langle d,\int \operatorname{co}\widetilde{F}\rangle = {\int }_{a}^{b}H\left( {d;F\left( t\right) }\right) {dt} = 0 \), and so \( \dim \int \operatorname{co}\widetilde{F} \leq \) \( K - 1 \) . By the inductive hypothesis, \( 0 \in \int \widetilde{F} \subset \int F \), which is the required conclusion. Step 3. We saw in Step 2 that if 0 is not in the relative interior of \( \int \operatorname{co}F \) , then a certain construction allows us to reduce the dimension and invoke the inductive hypothesis to conclude. The remaining case is that in which a \( \delta \) -neighborhood of 0 (in \( L \) ) lies in \( \int \operatorname{co}F \), for some \( \delta > 0 \) . Choose any nonzero vector \( {d}_{1} \) and define \( {\Phi }_{1} \) on \( {\mathbb{R}}^{n} \times X \) as follows: \[ {\Phi }_{1}\left( {x, s}\right) \mathrel{\text{:=}} {\int }_{a}^{b}\left\langle {{d}_{1}t + x, s\left( t\right) }\right\rangle {dt}. \] Given any \( x \) in \( {\mathbb{R}}^{n} \), express \( x \) as \( y + c \), where \( y \) lies in \( L \) and \( c \) in its orthogonal competent \( {L}^{ \bot } \) . Then \[ \mathop{\max }\limits_{{s \in S}}{\Phi }_{1}\left( {x, s}\right) = \mathop{\max }\limits_{{s \in S}}{\Phi }_{1}\left( {y, s}\right) \;\left( {\text{ since }\int \operatorname{co}F \subset L\text{ and }c \in {L}^{ \bot }}\right) \] \[ = \mathop{\max }\limits_{{s \in S}}{\int }_{a}^{b}\left\langle {{d}_{1}t + y, s\left( t\right) }\right\rangle {dt} \] \[ \geq \delta \parallel y\parallel + k \] for some \( k \) independent of \( x \) . It follows that the function \[ x \mapsto \mathop{\max }\limits_{{s \in S}}{\Phi }_{1}\left( {x, s}\right) \] has a minimum over \( {\mathbb{R}}^{n} \), say at \( {x}_{1} \) . The fact that \( S \) is weakly compact (Exercise 5.14) permits us to invoke the Minimax Theorem (see, for instance, Aubin (1993)) to deduce the existence of \( {s}_{0} \in S \) such that \[ \mathop{\min }\limits_{{x \in {\mathbb{R}}^{n}}}{\Phi }_{1}\left( {x,{s}_{0}}\right) = \mathop{\max }\limits_{{s \in S}}\mathop{\min }\limits_{{x \in {\mathbb{R}}^{n}}}{\Phi }_{1}\left( {x, s}\right) = \mathop{\min }\limits_{{x \in {\mathbb{R}}^{n}}}\mathop{\max }\limits_{{s \in S}}{\Phi }_{1}\left( {x, s}\right) \] \[ = \mathop{\max }\limits_{{s \in S}}{\Phi }_{1}\left( {{x}_{1}, s}\right) \] It follows that \( \left( {{x}_{1},{s}_{0}}\right) \) is a saddlepoint of \( {\Phi }_{1} \) relative to \( {\mathbb{R}}^{n} \times S \), so that the function \( {\Phi }_{1}\left( {\cdot ,{s}_{0}}\right) \) is minimized over \( {\mathbb{R}}^{n} \) at \( {x}_{1} \) . Setting the derivative equal to zero gives \( {\int }_{a}^{b}{s}_{0}\left( t\right) {dt} = 0 \) . The other saddlepoint inequality asserts that \( {\Phi }_{1}\left( {{x}_{1}, \cdot }\right) \) is maximized over \( S \) at \( {s}_{0} \) . But \[ \mathop{\max }\limits_{{s \in S}}{\Phi }_{1}\left( {{x}_{1}, \cdot }\right) = {\int }_{a}^{b}H\left( {{d}_{1}t + {x}_{1};F\left( t\right) }\right) {dt} = {\Phi }_{1}\left( {{x}_{1},{s}_{0}}\right) \] (by Exercise 5.4(b)), whence \[ H\left( {{d}_{1}t + {x}_{1};F\left( t\right) }\right) = \left\langle {{d}_{1}t + {x}_{1},{s}_{0}\left( t\right) }\right\rangle \text{ a.e. } \] Let us now define \[ {F}_{1}\left( t\right) \mathrel{\text{:=}} \left\{ {f \in F\left( t\right) : H\left( {{d}_{1}t + {x}_{1};F\left( t\right) }\right) = \left\langle {{d}_{1}t + {x}_{1}, f}\right\rangle }\right\} . \] We may confirm that \( {F}_{1} \) satisfies the same hypotheses as \( F \), and \( {s}_{0}\left( t\right) \in \) co \( {F}_{1}\left( t\right) \) a.e., whence \( 0 \in \int \operatorname{co}{F}_{1} \) . We now recommence either Step 2 or Step 3, with \( {F}_{1} \) in place of \( F \) . If Step 2 applies, we are finished, since \( \int {F}_{1} \subset \int F \) . If not, we perform Step 3 again, this time with a vector \( {d}_{2} \) linearly independent of \( {d}_{1} \) . This generates a new function \( {\Phi }_{2} \) admitting a saddlepoint at \( \left( {{x}_{2},{s}_{1}}\right) \), and a corresponding multifunction \( {F}_{2} \subset {F}_{1} \) all of whose points \( f \) satisfy \[ \left\langle {{d}_{2}t + {x}_{2}, f}\right\rangle = H\left( {{d}_{2}t + {x}_{2};{F}_{1}\left( t\right) }\right) \text{ a.e. } \] and such that \( 0 \in \int \operatorname{co}{F}_{2} \) . The process continues until either Step 2 has applied or else Step 3 has been performed \( n \) times. In the latter case, we will have defined a multifunction \( {F}_{n} \subset F \) such that \( 0 \in \int \operatorname{co}{F}_{n} \) and such that for almost all \( t \), every \( f \) in \( {F}_{n}\left( t\right) \) satisfies (we let \( {F}_{0} \) stand for \( F \) ): \[ \left\langle {{d}_{i}t + {x}_{i}, f}\right\rangle = H\left( {{d}_{i}t + {x}_{i};{F}_{i - 1}\left( t\right) }\right) \;\left( {i = 1,2,\ldots, n}\right) . \] In matrix terms, this may be written in the form \[ \left( {{Dt} + M}\right) f = \sum \left( t\right) \] where the \( n \times n \) matrix \( D \) is invertible (since its rows \( {d}_{i} \) are independent). But since \( {Dt} + M \) is invertible except for those (finitely many) \( t \) which are eigenvalues of \( - {D}^{-1}M \), it follows that \( {F}_{n}\left( t\right) \) is a singleton a.e. Consequently we have \( 0 \in \int \operatorname{co}{F}_{n} = \int {F}_{n} \subset \int F \) . 5.17. Exercise. Under the hypotheses of the theorem, we will prove the following corollary: Let \( \gamma \left( \cdot \right) \) be a measurable selection on \( \left\lbrack {a, b}\right\rbrack \) of \( \operatorname{co}F \) . Then there exists a sequence \( \left\{ {f}_{i}\right\} \) of measurable selections of \( F \) which converges weakly to \( \gamma \) . (a) For each positive integer \( N \), let \( a = : {t}_{0} < {t}_{1} < \cdots < {t}_{N} \mathrel{\text{:=}} b \) be a uniform partition of \( \left\lbrack {a, b}\right\rbrack \) ; invoke the theorem to deduce the existence of a selection \( {f}_{N} \) for \( F \) on \( \left\lbrack {a, b}\right\rbrack \) such that \[ {\int }_{{t}_{i}}^{{t}_{i + 1}}{f}_{N}\left( t\right) {dt} = {\int }_{{t}_{i}}^{{t}_{i + 1}}\gamma \left( t\right) {dt}\;\left( {i = 0,1,\ldots, N - 1}\right) . \] (b) Prove that the sequence \( \left\{ {f}_{N}\right\} \) admits a subsequence \( \left\{ {f}_{{N}_{i}}\right\} \) converging weakly to a limit \( \zeta \) (see Exercise 5.14). Then show that \( \zeta \) must be \( \gamma \) . (Hint. Study the convergence of \( \left\langle {{f}_{{N}_{i}}, g}\right\rangle \) when \( g \) is a smooth function, using integration by parts.) We now return to the analysis of \( {\partial }_{L}f \), where \[ f\left( x\right) = {\int }_{a}^{b}\varphi \left( {x\left( t\right) }\right) {dt} \] The hypotheses remain those of Theorem 5.10. 5.18. Theorem. The limiting subdifferential and the generalized gradient of \( f \) coincide, and we have \[ {\partial }_{L}f\left( x\right) = {\partial }_{C}f\left( x\right) = \left\{ {\zeta \in {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack : \zeta \left( t\right) \in {\partial }_{C}\varphi \left( {x\left( t\right) }\right) \text{ a.e. }}\right\} . \] Proof. We set \[ \Lambda \mathrel{\text{:=}} \left\{ {\zeta = \mathop{\operatorname{w-lim}}\limits_{{i \rightarrow \infty }}{\zeta }_{i} : {\zeta }_{i}\left( t\right) \in {\partial }_{L}\varphi \left( {x\left( t\right) }\right) \text{ a.e. }}\right\} \] The following three relations will be established in order: \[ {\partial }_{C}f\left( x\right) \subset \left\{ {\zeta : \zeta \left( t\right) \in {\partial }_{C}\varphi \left( {x\left( t\right) }\right) \text{ a.e. }}\right\} \subset \Lambda \subset {\partial }_{L}f\left( x\right) . \] This evidently implies the theorem, since \( {\partial }_{L}f\left( x\right) \subset {\partial }_{C}f\left( x\right) \) . The set \[ \left\{ {\zeta : \zeta \left( t\right) \in {\partial }_{C}\varphi \left( {x\left( t\right) }\right) \text{ a.e. }}\right\} \] is convex and weakly compact in \( X \) (see Exercise 5.14), as is \( {\partial }_{C}f\left( x\right) \) . The first relation can therefore be proven via support functions; it amounts to showing that for any \( v \in X \), we have \[ {f}^{ \circ }\left( {x;v}\right) \leq \max \left\{ {\langle v,\zeta \rangle : \zeta \left( t\right) \in {\partial }_{C}\varphi \left( {x\left( t\right) }\right) \text{ a.e. }}\right\} \] Let \( \left\{ {x}_{i}\right\} \) be a sequence in \( X \) converging to \( x \) and \( \left\{
1012_(GTM178)Nonsmooth Analysis and Control Theory
57
rtial }_{C}f\left( x\right) \subset \left\{ {\zeta : \zeta \left( t\right) \in {\partial }_{C}\varphi \left( {x\left( t\right) }\right) \text{ a.e. }}\right\} \subset \Lambda \subset {\partial }_{L}f\left( x\right) . \] This evidently implies the theorem, since \( {\partial }_{L}f\left( x\right) \subset {\partial }_{C}f\left( x\right) \) . The set \[ \left\{ {\zeta : \zeta \left( t\right) \in {\partial }_{C}\varphi \left( {x\left( t\right) }\right) \text{ a.e. }}\right\} \] is convex and weakly compact in \( X \) (see Exercise 5.14), as is \( {\partial }_{C}f\left( x\right) \) . The first relation can therefore be proven via support functions; it amounts to showing that for any \( v \in X \), we have \[ {f}^{ \circ }\left( {x;v}\right) \leq \max \left\{ {\langle v,\zeta \rangle : \zeta \left( t\right) \in {\partial }_{C}\varphi \left( {x\left( t\right) }\right) \text{ a.e. }}\right\} \] Let \( \left\{ {x}_{i}\right\} \) be a sequence in \( X \) converging to \( x \) and \( \left\{ {\lambda }_{i}\right\} \) a positive sequence converging to 0 such that \[ {f}^{ \circ }\left( {x;v}\right) = \mathop{\lim }\limits_{{i \rightarrow \infty }}\frac{f\left( {{x}_{i} + {\lambda }_{i}v}\right) - f\left( {x}_{i}\right) }{{\lambda }_{i}}. \] The limit can be written \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}{\int }_{a}^{b}\frac{\varphi \left( {{x}_{i}\left( t\right) + {\lambda }_{i}v\left( t\right) }\right) - \varphi \left( {{x}_{i}\left( t\right) }\right) }{{\lambda }_{i}} \] \[ \leq {\int }_{a}^{b}\mathop{\limsup }\limits_{{i \rightarrow \infty }}\frac{\varphi \left( {{x}_{i}\left( t\right) + {\lambda }_{i}v\left( t\right) }\right) - \varphi \left( {{x}_{i}\left( t\right) }\right) }{{\lambda }_{i}}{dt} \] (Why has Fatou's Lemma been applicable?) \[ \leq {\int }_{a}^{b}{\varphi }^{ \circ }\left( {x\left( t\right) ;v\left( t\right) }\right) {dt} \] (since \( {x}_{i}\left( t\right) \rightarrow x\left( t\right) \) a.e.). By Exercise 5.4(b) there exists a measurable selection \( \zeta \left( \cdot \right) \) of \( {\partial }_{C}\varphi \left( {x\left( \cdot \right) }\right) \) such that \( \langle \zeta \left( t\right), v\left( t\right) \rangle = {\varphi }^{ \circ }\left( {x\left( t\right) ;v\left( t\right) }\right) \) a.e. Then \[ {\int }_{a}^{b}{\varphi }^{ \circ }\left( {x\left( t\right) ;v\left( t\right) }\right) {dt} = \langle \zeta, v\rangle \] and the first of the three relations follows. The second relation follows immediately from Exercise 5.17, since \[ {\partial }_{C}\varphi \left( {x\left( t\right) }\right) = \operatorname{co}{\partial }_{L}\varphi \left( {x\left( t\right) }\right) \] There remains the third. Let \( \zeta = \mathrm{w} - \mathop{\lim }\limits_{{i \rightarrow \infty }}{\zeta }_{i} \), where \( {\zeta }_{i}\left( t\right) \in {\partial }_{L}\varphi \left( {x\left( t\right) }\right) \) a.e. We wish to prove that \( \zeta \in {\partial }_{L}f\left( x\right) \) . As observed in Problem 1.11.28, there is a countable set \( C \) in \( {\mathbb{R}}^{n} \times {\mathbb{R}}^{n} \) (not depending on \( t \) ) of points \( \left( {\xi, y}\right) \) having \( \xi \in {\partial }_{P}\varphi \left( y\right) \) such that for each \( t \) we have \[ {\partial }_{L}\varphi \left( {x\left( t\right) }\right) = \left\{ {\mathop{\lim }\limits_{{k \rightarrow \infty }}{\xi }_{k} : \left( {{\xi }_{k},{y}_{k}}\right) \in C,\mathop{\lim }\limits_{{k \rightarrow \infty }}{y}_{k} = x\left( t\right) }\right\} . \] For a given enumeration \( \left\{ \left( {{\xi }_{j},{y}_{j}}\right) \right\} \) of \( C \), let \( {j}_{i}\left( t\right) \) be the first \( j \) such that \[ \begin{Vmatrix}{{\xi }_{j} - {\zeta }_{i}\left( t\right) }\end{Vmatrix} < {i}^{-1}\;\text{ and }\;\begin{Vmatrix}{{y}_{j} - x\left( t\right) }\end{Vmatrix} < {i}^{-1}. \] Let us set \[ {\widetilde{x}}_{i}\left( t\right) \mathrel{\text{:=}} {y}_{{j}_{i}\left( t\right) },\;{\widetilde{\zeta }}_{i}\left( t\right) \mathrel{\text{:=}} {\xi }_{{j}_{i}\left( t\right) }. \] Then \( {\widetilde{x}}_{i}\left( \cdot \right) \) and \( {\widetilde{\zeta }}_{i}\left( \cdot \right) \) are measurable, and \[ {\widetilde{\zeta }}_{i}\left( t\right) \in {\partial }_{P}\varphi \left( {{\widetilde{x}}_{i}\left( t\right) }\right) \text{a.e.} \] By Proposition 5.12, there exist \( {x}_{i}^{\prime } \) and \( {\zeta }_{i}^{\prime } \) in \( X \) such that \[ \begin{Vmatrix}{{x}_{i}^{\prime } - {\widetilde{x}}_{i}}\end{Vmatrix} < {i}^{-1},\;\begin{Vmatrix}{{\zeta }_{i}^{\prime } - {\widetilde{\zeta }}_{i}}\end{Vmatrix} < {i}^{-1},\;{\zeta }_{i}^{\prime } \in {\partial }_{P}f\left( {x}_{i}^{\prime }\right) . \] It follows that \( \left\{ {\zeta }_{i}^{\prime }\right\} \) converges weakly to \( \zeta \), and \( \left\{ {x}_{i}^{\prime }\right\} \) strongly to \( x \) . Of course, \( f\left( {x}_{i}^{\prime }\right) \rightarrow f\left( x\right) \) since \( f \) is Lipschitz. But then \( \zeta \in {\partial }_{L}f\left( x\right) \) by definition, as required. 5.19. Exercise. It is not the case that \( {\partial }_{L}f\left( x\right) \) is convex for every Lipschitz functional \( f \) on \( X \mathrel{\text{:=}} {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) . Verify this statement with \( n = 1 \) for the functional \[ f\left( x\right) \mathrel{\text{:=}} - \left| {{\int }_{0}^{1}x\left( t\right) {dt}}\right| \] ## A Problem in the Calculus of Variations Recall that a function \( x : \left\lbrack {a, b}\right\rbrack \rightarrow {\mathbb{R}}^{n} \) is said to be absolutely continuous if it can be expressed in the form \[ x\left( t\right) = {x}_{0} + {\int }_{a}^{t}v\left( s\right) {ds} \] for some integrable function \( v \) ; we then have \( \dot{x}\left( t\right) \mathrel{\text{:=}} d/{dtx}\left( t\right) = v\left( t\right) \) a.e. We now consider the variational problem of minimizing the functional \[ \ell \left( {x\left( b\right) }\right) + {\int }_{a}^{b}\varphi \left( {x\left( t\right) ,\dot{x}\left( t\right) }\right) {dt} \] over those absolutely continuous functions \( x : \left\lbrack {a, b}\right\rbrack \rightarrow {\mathbb{R}}^{n} \) which satisfy \( x\left( 0\right) = {x}_{0} \) . Here the "givens" of the problem are the point \( {x}_{0} \), the interval \( \left\lbrack {a, b}\right\rbrack \), the function \( \ell : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) (assumed locally Lipschitz) and the function \( \varphi : {\mathbb{R}}^{n} \times {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) (assumed globally Lipschitz). Our goal is to derive necessary conditions for optimality. Let us rephrase the problem so as to make the relevance of earlier results more apparent. We define the following subset \( A \) of \( {L}_{2n}^{2}\left\lbrack {a, b}\right\rbrack \) : \[ A \mathrel{\text{:=}} \left\{ {\left( {u, v}\right) \in {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \times {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack : u\left( t\right) = {x}_{0} + {\int }_{a}^{t}v\left( s\right) {ds}, t \in \left\lbrack {a, b}\right\rbrack }\right\} . \] We note that \( A \) is closed and convex. Now define \( {f}_{1},{f}_{2} \) on \( {L}_{2n}^{2}\left\lbrack {a, b}\right\rbrack \) as follows: \[ {f}_{1}\left( {u, v}\right) \mathrel{\text{:=}} \ell \left( {{x}_{0} + {\int }_{a}^{b}v\left( t\right) {dt}}\right) ,\;{f}_{2}\left( {u, v}\right) \mathrel{\text{:=}} {\int }_{a}^{b}\varphi \left( {u\left( t\right), v\left( t\right) }\right) {dt}. \] In these terms, our problem becomes: \[ \operatorname{minimize}\left\{ {{f}_{1}\left( {u, v}\right) + {f}_{2}\left( {u, v}\right) : \left( {u, v}\right) \in A.}\right\} \] We have characterized \( {\partial }_{L}{f}_{2} \) in Theorem 5.18; as for \( {\partial }_{L}{f}_{1} \), we have: 5.20. Exercise. The functional \( {f}_{1} \) is locally Lipschitz on \( {L}_{2n}^{2}\left\lbrack {a, b}\right\rbrack \) . If \( \left( {\theta ,\zeta }\right) \in {\partial }_{L}{f}_{1}\left( {u, v}\right) \), then \( \theta = 0 \), and for some \( {\zeta }_{0} \in {\partial }_{L}\ell \left( {{x}_{0} + {\int }_{a}^{b}v\left( t\right) {dt}}\right. \) ), we have \( \zeta \left( t\right) = {\zeta }_{0} \) a.e. The final component we will need is the following version of a celebrated result of the nineteenth century. 5.21. Proposition (Dubois-Reymond Lemma). Let \( \left( {\zeta ,\xi }\right) \in {N}_{A}^{P}\left( {u, v}\right) \) . Then \[ \xi \left( t\right) = - {\int }_{t}^{b}\zeta \left( s\right) {ds},\;a \leq t \leq b, \] so that \( \xi \) is absolutely continuous and satisfies \( \xi \left( b\right) = 0 \) . Proof. We remark that since \( A \) is convex, the normal cones \( {N}_{A}^{P},{N}_{A}^{L} \), and \( {N}_{A}^{C} \) all coincide. The normal vector \( \left( {\zeta ,\xi }\right) \) satisfies \[ \left\langle {\left( {\zeta ,\xi }\right) ,\left( {{u}^{\prime },{v}^{\prime }}\right) }\right\rangle \leq \langle \left( {\zeta ,\xi }\right) ,\left( {u, v}\right) \rangle \forall \left( {{u}^{\prime },{v}^{\prime }}\right) \in A, \] which translates as \[ {\int }_{a}^{b}\left\{ {\zeta \left( t\right) \cdot \left( {{u}^{\prime }\left( t\right) - u\left( t\right) }\right) + \xi \left( t\right) \cdot \left( {{v}^{\prime }\left( t\right) - v\left( t\right) }\right) }\right\} {dt} \leq 0. \] Bearing in mind that \( {u}^{\prime }\left( t\right) - u\left( t\right) = {\int }_{a}^{t}\left( {{v}^{\prime }\left( s\right) - v\left( s\right) }\right) {ds} \), we use integration by parts to derive from this last inequality the following one: \[ {\int }_{a}^{b}\left\lbrack {\xi \left( t\right) + {\int }_{t}^{b}\zeta \left( s\right) {ds}}\right\rbrack \cdot \left\lbrack {{v}^{\prime }\left( t\right) - v\left( t\right) }\right\rbrack {dt} \leq 0\forall {v}^{\prime } \in {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack . \] Since \( {v}^{\prime } - v \) is an arbitrary integrable function, the required conclusion follows immediately. Combining all the ingredients, we arrive at a nonsmooth version of the famous Euler equation in the calculus of v
1012_(GTM178)Nonsmooth Analysis and Control Theory
58
es as \[ {\int }_{a}^{b}\left\{ {\zeta \left( t\right) \cdot \left( {{u}^{\prime }\left( t\right) - u\left( t\right) }\right) + \xi \left( t\right) \cdot \left( {{v}^{\prime }\left( t\right) - v\left( t\right) }\right) }\right\} {dt} \leq 0. \] Bearing in mind that \( {u}^{\prime }\left( t\right) - u\left( t\right) = {\int }_{a}^{t}\left( {{v}^{\prime }\left( s\right) - v\left( s\right) }\right) {ds} \), we use integration by parts to derive from this last inequality the following one: \[ {\int }_{a}^{b}\left\lbrack {\xi \left( t\right) + {\int }_{t}^{b}\zeta \left( s\right) {ds}}\right\rbrack \cdot \left\lbrack {{v}^{\prime }\left( t\right) - v\left( t\right) }\right\rbrack {dt} \leq 0\forall {v}^{\prime } \in {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack . \] Since \( {v}^{\prime } - v \) is an arbitrary integrable function, the required conclusion follows immediately. Combining all the ingredients, we arrive at a nonsmooth version of the famous Euler equation in the calculus of variations: 5.22. Theorem (Euler Inclusion). If \( x \) solves the variational problem, then there exists an absolutely continuous function \( p \) satisfying \[ \left( {\dot{p}\left( t\right), p\left( t\right) }\right) \in {\partial }_{C}\varphi \left( {x\left( t\right) ,\dot{x}\left( t\right) }\right) \text{a.e.,} \] \[ - p\left( b\right) \in {\partial }_{L}\ell \left( {x\left( b\right) }\right) . \] Proof. The pair \( \left( {x,\dot{x}}\right) \) minimizes the function \[ {f}_{1}\left( {u, v}\right) + {f}_{2}\left( {u, v}\right) + {I}_{A}\left( {u, v}\right) \] over \( {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \times {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) . Thus \[ \left( {0,0}\right) \in {\partial }_{L}\left\{ {{f}_{1} + {f}_{2} + {I}_{A}}\right\} \left( {x,\dot{x}}\right) \] \[ \subset {\partial }_{L}{f}_{1}\left( {x,\dot{x}}\right) + {\partial }_{L}{f}_{2}\left( {x,\dot{x}}\right) + {N}_{A}^{C}\left( {x,\dot{x}}\right) \] by Proposition 1.10.1. Combining Exercise 5.20, Theorem 5.18, and Proposition 5.21, we deduce the existence of a point \( {\zeta }_{0} \in {\partial }_{L}\ell \left( {x\left( b\right) }\right) \) and an element \( \zeta \) of \( {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) such that \[ \left( {-\zeta \left( t\right) , - {\zeta }_{0} + {\int }_{t}^{b}\zeta \left( s\right) {ds}}\right) \in {\partial }_{C}\varphi \left( {x\left( t\right) ,\dot{x}\left( t\right) }\right) \text{ a.e. } \] The statement of the theorem follows upon setting \( p\left( t\right) \mathrel{\text{:=}} - {\zeta }_{0} + {\int }_{t}^{b}\zeta \left( s\right) {ds} \) . 5.23. Exercise. When \( \varphi \) is \( {C}^{1} \), a solution \( x \) to the variational prob- lem satisfies the classical Euler equation \[ \frac{d}{dt}\left\{ {{\varphi }_{v}\left( {x\left( t\right) ,\dot{x}\left( t\right) }\right) }\right\} = {\varphi }_{u}\left( {x\left( t\right) ,\dot{x}\left( t\right) }\right) \text{ a.e.,} \] and the function \( t \mapsto {\varphi }_{v}\left( {x\left( t\right) ,\dot{x}\left( t\right) }\right) \) is continuous. We remark that the last conclusion of this exercise is known as the first Erdmann condition in the classical calculus of variations. The relation \( - p\left( b\right) = \nabla \ell \left( {x\left( b\right) }\right) \) is the transversality condition. ## A Weak Sequential Compactness Theorem We conclude this section with a technical result of some importance in the next chapter. It concerns a multifunction \( E \) mapping \( \mathbb{R} \times {\mathbb{R}}^{n} \) to the closed convex subsets of a given compact set \( {E}_{0} \) in \( {\mathbb{R}}^{n} \) . We assume that \( E \) is graph-closed. 5.24. Theorem. Let \( \left\{ {v}_{i}\right\} \) be a sequence in \( {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) such that \[ {v}_{i}\left( t\right) \in E\left( {{\tau }_{i}\left( t\right) ,{u}_{i}\left( t\right) }\right) + {r}_{i}\left( t\right) \bar{B}\text{ a.e.,}t \in \left\lbrack {a, b}\right\rbrack , \] where the sequence of measurable functions \( \left\{ {{\tau }_{i}\left( \cdot \right) ,{u}_{i}\left( \cdot \right) }\right\} \) converges a.e. to \( \left( {t,{u}_{0}\left( t\right) }\right) \), and where the nonnegative measurable functions \( \left\{ {r}_{i}\right\} \) converge to 0 in \( {L}_{1}^{2}\left\lbrack {a, b}\right\rbrack \) . Then there exists a subsequence \( \left\{ {v}_{{i}_{j}}\right\} \) of \( \left\{ {v}_{i}\right\} \) which converges weakly in \( {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) to a limit \( {v}_{0}\left( \cdot \right) \) which satisfies \[ {v}_{0}\left( t\right) \in E\left( {t,{u}_{0}\left( t\right) }\right) \text{ a.e.,}t \in \left\lbrack {a, b}\right\rbrack . \] Proof. The hypotheses evidently imply that the sequence \( \left\{ {v}_{i}\right\} \) is bounded in \( {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) . Invoking weak compactness, we know that a subsequence \( \left\{ {v}_{{i}_{j}}\right\} \) converges weakly to some limit \( {v}_{0} \) ; there remains only to prove that \( {v}_{0}\left( t\right) \) belongs to \( E\left( {t,{u}_{0}\left( t\right) }\right) \) a.e. for \( t \in \left\lbrack {a, b}\right\rbrack \) . We define \( h : \mathbb{R} \times {\mathbb{R}}^{n} \times {\mathbb{R}}^{n} \rightarrow ( - \infty ,\infty \rbrack \) as follows: \[ h\left( {t, u, p}\right) \mathrel{\text{:=}} \min \{ \langle p, v\rangle : v \in E\left( {t, u}\right) \} . \] Since \( E \) is convex-valued, a point \( v \) lies in \( E\left( {t, u}\right) \) iff \( \langle p, v\rangle \geq h\left( {t, u, p}\right) \) \( \forall p \in {\mathbb{R}}^{n} \) (by the Separation Theorem). In view of the inclusion satisfied by \( {v}_{i} \), we also have, for any fixed \( p \in {\mathbb{R}}^{n} \), the following inequality: \[ \left\langle {p,{v}_{i}\left( t\right) }\right\rangle + {r}_{i}\left( t\right) \parallel p\parallel \geq h\left( {{\tau }_{i}\left( t\right) ,{u}_{i}\left( t\right), p}\right) \text{ a.e. } \] Because the multifunction \( E \) has closed graph, it follows readily that the function \( \left( {t, u}\right) \mapsto h\left( {t, u, p}\right) \) is lower semicontinuous, and hence that the function \[ t \mapsto h\left( {{\tau }_{i}\left( t\right) ,{u}_{i}\left( t\right), p}\right) \] is measurable (Exercise 5.2(g)). Now let \( A \) be any measurable subset of \( \left\lbrack {a, b}\right\rbrack \) . We have \[ {\int }_{A}\left\{ {\left\langle {p,{v}_{{i}_{j}}\left( t\right) }\right\rangle + \parallel p\parallel {r}_{{i}_{j}}\left( t\right) - h\left( {{\tau }_{{i}_{j}}\left( t\right) ,{u}_{{i}_{j}}\left( t\right), p}\right) }\right\} {dt} \geq 0, \] since the integrand is nonnegative almost everywhere. As \( j \rightarrow \infty \), we know \[ {\int }_{A}\left\langle {p,{v}_{{i}_{j}}\left( t\right) }\right\rangle {dt} \rightarrow {\int }_{A}\left\langle {p,{v}_{0}\left( t\right) }\right\rangle {dt} \] by weak convergence, and of course \( {\int }_{A}{r}_{{i}_{j}}\left( t\right) {dt} \rightarrow 0 \) . In light of this, and calling upon Fatou's Lemma (Why does it apply?), we deduce \[ {\int }_{A}\left\langle {p,{v}_{0}\left( t\right) }\right\rangle {dt} - {\int }_{A}\mathop{\liminf }\limits_{{j \rightarrow \infty }}h\left( {{\tau }_{{i}_{j}}\left( t\right) ,{u}_{{i}_{j}}\left( t\right), p}\right) {dt} \geq 0. \] Because \( h\left( {t, u, p}\right) \) is lower semicontinuous in \( \left( {t, u}\right) \), this implies \[ {\int }_{A}\left\{ {\left\langle {p,{v}_{0}\left( t\right) }\right\rangle - h\left( {t,{u}_{0}\left( t\right), p}\right) }\right\} {dt} \geq 0. \] Since \( A \) is arbitrary, we must then have \[ \left\langle {p,{v}_{0}\left( t\right) }\right\rangle \geq h\left( {t,{u}_{0}\left( t\right), p}\right) \text{ a.e.,}t \in \left\lbrack {a, b}\right\rbrack . \] (7) Now let \( \left\{ {p}_{i}\right\} \) be a countable dense subset of \( {\mathbb{R}}^{n} \) . Then (7) holds for each \( p = {p}_{i} \), with the exception of \( t \) in an exceptional set \( {\Omega }_{i} \) of measure 0 in \( \left\lbrack {a, b}\right\rbrack \) . Let \( \Omega \mathrel{\text{:=}} \mathop{\bigcup }\limits_{i}{\Omega }_{i} \) . Then for all \( t \notin \Omega \) ,(7) holds for all \( p \in \left\{ {p}_{i}\right\} \) . But both sides of (7) define continuous functions of \( p \), for given \( t \) . (To see that \( h \) is continuous in \( p \), it suffices to observe that it is concave and finite as a function of \( p \) .) Therefore for any \( t \notin \Omega \) ,(7) must in fact hold for all \( p \in {\mathbb{R}}^{n} \) , which is equivalent to \[ {v}_{0}\left( t\right) \in E\left( {t,{u}_{0}\left( t\right) }\right) \forall t \in \Omega . \] Since \( \Omega \) has measure 0, the proof is complete. We remark that the convexity of the values of \( E \) plays a crucial role in the theorem; Problem 7.19 will make this clear. ## 6 Tangents and Interiors In this section we study certain properties of sets related to tangency and interiors. The discussion will be restricted to subsets of \( {\mathbb{R}}^{n} \), and one of the central issues is the following: If at a point \( x \in S \) the set \( S \) admits a "large" set of tangents, does it follow that \( S \) is "substantial" near \( x \) ? For example, can we assert that the interior of \( S \) is nonempty, and that \( x \in \operatorname{cl}\operatorname{int}S \) ? In answering such questions, the tangent cone \( {T}_{S}^{C}\left( x\right) \) turns out to be more useful than \( {T}_{S}^{B}\left( x\right) \) . Furthermore, the property that int \( {T}_{S}^{C}\left( x\right) \neq \varnothing \) will serve to identify a class of sets having important properties which play a role, for example, in the theory of equilibria and in constructing feedback controls. Let \( S \) be a closed nonempty subset of \( {\mathbb{R}}^{n} \), and let \( x \in S \) . We will say that \( S \) is wedged at \( x \) provided that \( \operatorname{int}{T}_{S}^{C}\left( x\right) \neq \varnothing \) . Since \( {T}_{S}^{C}\left( x\right)
1012_(GTM178)Nonsmooth Analysis and Control Theory
59
s of \( {\mathbb{R}}^{n} \), and one of the central issues is the following: If at a point \( x \in S \) the set \( S \) admits a "large" set of tangents, does it follow that \( S \) is "substantial" near \( x \) ? For example, can we assert that the interior of \( S \) is nonempty, and that \( x \in \operatorname{cl}\operatorname{int}S \) ? In answering such questions, the tangent cone \( {T}_{S}^{C}\left( x\right) \) turns out to be more useful than \( {T}_{S}^{B}\left( x\right) \) . Furthermore, the property that int \( {T}_{S}^{C}\left( x\right) \neq \varnothing \) will serve to identify a class of sets having important properties which play a role, for example, in the theory of equilibria and in constructing feedback controls. Let \( S \) be a closed nonempty subset of \( {\mathbb{R}}^{n} \), and let \( x \in S \) . We will say that \( S \) is wedged at \( x \) provided that \( \operatorname{int}{T}_{S}^{C}\left( x\right) \neq \varnothing \) . Since \( {T}_{S}^{C}\left( x\right) = {\mathbb{R}}^{n} \) for any \( x \in \operatorname{int}S \), this property is at issue only for points \( x \) lying in bdry \( S \) . 6.1. Exercise. Which of the sets described in Exercise 2.5.6 is wedged at 0 ? We have already observed in Problem 2.9.10 the following fact: 6.2. Proposition. Let \( S = \left\{ {{x}^{\prime } : f\left( {x}^{\prime }\right) \leq 0}\right\} \), where \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is Lipschitz near \( x \) and \( 0 \notin \partial f\left( x\right) \) . Then \( S \) is wedged at \( x \) . 6.3. Exercise. (a) If \( S \) is given as in Proposition 6.2, then int \( S \neq \varnothing \), and \( x \in \) cl(int \( S \) ). (Hint. Consider Problem 2.9.6.) (b) By taking \( n = 2 \) and \( f\left( {x, y}\right) = \left| x\right| - \left| y\right| \), show that Proposition 6.2 fails when \( \partial f \) is replaced by \( {\partial }_{L}f \) . 6.4. Theorem. A vector \( v \in {\mathbb{R}}^{n} \) belongs to \( \operatorname{int}{T}_{S}^{C}\left( x\right) \) iff there exists \( \varepsilon > 0 \) such that \[ y \in x + {\varepsilon B},\;w \in v + {\varepsilon B},\;t \in \lbrack 0,\varepsilon ) \Rightarrow {d}_{S}\left( {y + {tw}}\right) \leq {d}_{S}\left( y\right) . \] (1) Proof. Let \( v \in \operatorname{int}{T}_{S}^{C}\left( x\right) \) . If \( v = 0 \), then \( {T}_{S}^{C}\left( x\right) = {\mathbb{R}}^{n} \), so that by polarity (see Proposition 2.5.4) we have \( {N}_{S}^{C}\left( x\right) = \{ 0\} \) . But \( {N}_{S}^{L}\left( x\right) \subset {N}_{S}^{C}\left( x\right) \) is nontrivial when \( x \in \operatorname{bdry}S \) (Exercise 2.8.5), so that \( x \in \operatorname{int}S \) necessarily. But in that case property (1) evidently holds for a suitably small \( \varepsilon > 0 \) . So let us suppose \( 0 \neq v \in \operatorname{int}{T}_{S}^{C}\left( x\right) \) . Then by polarity there exists \( \delta > 0 \) such that \[ \langle v,\zeta \rangle \leq - \delta \parallel \zeta \parallel \forall \zeta \in {N}_{S}^{C}\left( x\right) . \] If (1) fails to hold for any \( \varepsilon > 0 \), then there exist sequences \( \left\{ {y}_{i}\right\} ,\left\{ {w}_{i}\right\} \) converging to \( x \) and \( v \), respectively, and a positive sequence \( \left\{ {t}_{i}\right\} \) decreasing to 0 such that \[ {d}_{S}\left( {{y}_{i} + {t}_{i}{w}_{i}}\right) - {d}_{S}\left( {y}_{i}\right) > 0. \] We invoke the Proximal Mean Value Theorem to deduce the existence of \( {z}_{i} \) converging to \( x \) and \( {\zeta }_{i} \in {\partial }_{P}{d}_{S}\left( {z}_{i}\right) \) such that \( \left\langle {{\zeta }_{i},{w}_{i}}\right\rangle > 0 \) . Take a subsequence if necessary to have \( {\zeta }_{i}/\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \) converge to a limit \( \zeta \in {\partial }_{L}{d}_{S}\left( x\right) \subset \) \( {N}_{S}^{C}\left( x\right) \) . We have \[ - \delta = - \delta \parallel \zeta \parallel \geq \langle v,\zeta \rangle = \mathop{\lim }\limits_{{i \rightarrow \infty }}\left\langle {{w}_{i},{\zeta }_{i}}\right\rangle /\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \geq 0, \] a contradiction which establishes the necessity of (1). As for the sufficiency, it is evident that (1) implies \[ {d}_{S}^{ \circ }\left( {x;w}\right) \leq 0\forall w \in v + {\varepsilon B} \] whence \( B\left( {v;\varepsilon }\right) \subset {T}_{S}^{C}\left( x\right) \) and \( v \in \operatorname{int}{T}_{S}^{C}\left( x\right) \) . Let us agree to call the following set \( W\left( {v;\varepsilon }\right) \) a wedge (of axis \( v \) and radius \( \varepsilon ) \) : \[ W\left( {v;\varepsilon }\right) \mathrel{\text{:=}} \{ {tw} : t \in \lbrack 0,\varepsilon ), w \in v + {\varepsilon B}\} . \] The first of the several immediate consequences of the theorem given below explains our use of the term "wedged" for the condition \( 0 \in \operatorname{int}{T}_{S}^{C}\left( x\right) \), and reveals it to be of a local and uniform nature. ## 6.5. Exercise. (a) \( S \) is wedged at \( x \) iff there exists a wedge \( W\left( {v;\varepsilon }\right) \) such that \( y + W\left( {v;\varepsilon }\right) \subset S\forall y \in S \cap B\left( {x;\varepsilon }\right) . \) (b) If \( S \) is wedged at \( x \), then int \( S \neq \varnothing \) and \( x \in \operatorname{cl}\left( {\operatorname{int}S}\right) \) . If \( S \) is wedged at each of its points, then \( S = \operatorname{cl}\left( {\operatorname{int}S}\right) \) . (c) If \( v \in \operatorname{int}{T}_{S}^{C}\left( x\right) \), then \( v \in {T}_{S}^{C}\left( {x}^{\prime }\right) \) for all \( {x}^{\prime } \) near \( x \) . (d) If \( {T}_{S}^{C}\left( x\right) = {\mathbb{R}}^{n} \), then \( x \in \operatorname{int}S \) . (e) Let \( n = 2 \) and set \( S \) equal to \[ \{ \left( {x, y}\right) : \parallel \left( {x, y}\right) - \left( {1,0}\right) \parallel \leq 1\} \cup \{ \left( {x, y}\right) : \parallel \left( {x, y}\right) - \left( {-1,0}\right) \parallel \leq 1\} . \] Show that \( {T}_{S}^{B}\left( {0,0}\right) = {\mathbb{R}}^{2} \), yet \( \left( {0,0}\right) \notin \operatorname{int}S \) . What is \( {T}_{S}^{C}\left( {0,0}\right) \) ? ## Lower Semicontinuity of \( {T}_{S}^{C}\left( \cdot \right) \) A multifunction \( \Gamma : X \rightarrow X \) is said to be lower semicontinuous at \( x \) if for any given \( v \in \Gamma \left( x\right) \) and \( \varepsilon > 0 \), there exists \( \delta > 0 \) such that \[ {x}^{\prime } \in \operatorname{dom}\Gamma ,\;{x}^{\prime } \in x + {\delta B} \Rightarrow v \in \Gamma \left( {x}^{\prime }\right) + {\varepsilon B}. \] 6.6. Exercise. (a) Let \( \Delta \mathrel{\text{:=}} \operatorname{dom}\Gamma \) . Then \( \Gamma \) is lower semicontinuous at \( x \in \Delta \) iff for every \( v \in \Gamma \left( x\right) \) we have \[ \mathop{\limsup }\limits_{{{x}^{\prime } \triangleq x}}d\left( {v,\Gamma \left( {x}^{\prime }\right) }\right) = 0. \] (b) In Exercise 2.5.6, determine in which cases \( {T}_{S}^{C}\left( \cdot \right) \) or \( {T}_{S}^{B}\left( \cdot \right) \) is lower semicontinuous at the origin. The lower semicontinuity of the tangent cone is a very useful property in certain contexts, and it is notable that wedged sets automatically possess it. 6.7. Proposition. If \( S \) is wedged at \( x \), then \( {T}_{S}^{C}\left( \cdot \right) \) is lower semicontinuous at \( x \) . Proof. Given \( v \in {T}_{S}^{C}\left( x\right) \), we must verify (see Exercise 6.6) \[ \mathop{\limsup }\limits_{{{x}^{\prime }\overset{S}{ \rightarrow }x}}d\left( {v,{T}_{S}^{C}\left( {x}^{\prime }\right) }\right) = 0. \] Since \( {T}_{S}^{C}\left( x\right) \) is closed convex with nonempty interior, it is the closure of its interior. In view of this fact, it clearly suffices to verify the preceding condition for \( v \in \operatorname{int}{T}_{S}^{C}\left( x\right) \) . But for such \( v \), it is immediate from Exercise 6.5(c). The following reflects a general fact about closed convex cone-valued mul-tifunctions which are in polarity. 6.8. Proposition. \( {T}_{S}^{C}\left( \cdot \right) \) is lower semicontinuous at \( x \) iff \( {N}_{S}^{C}\left( \cdot \right) \) is graph-closed at \( x \) . Proof. Let \( {T}_{S}^{C}\left( \cdot \right) \) be lower semicontinuous at \( x \), and let \( {\zeta }_{i} \in {N}_{S}^{C}\left( {x}_{i}\right) \), where \( {x}_{i} \rightarrow x,{\zeta }_{i} \rightarrow \zeta \) . We wish to deduce \( \zeta \in {N}_{S}^{C}\left( x\right) \) . Without loss of generality, we can suppose \( \parallel \zeta \parallel = 1 \) . If \( \zeta \notin {N}_{S}^{C}\left( x\right) \), then there exists a vector \( v \) separating \( \zeta \) from \( {N}_{S}^{C}\left( x\right) \) : For some \( \delta > 0 \) we have \[ \langle v,\xi \rangle \leq 0 < \delta = \langle v,\zeta \rangle \forall \xi \in {N}_{S}^{C}\left( x\right) . \] It follows that \( v \in {N}_{S}^{C}{\left( x\right) }^{ \circ } = {T}_{S}^{C}\left( x\right) \), and that (by lower semicontinuity) \( v \in {T}_{S}^{C}\left( {x}_{i}\right) + \left( {\delta /2}\right) B \) for all \( i \) sufficiently large. Then, for some vector \( {u}_{i} \in B \), and for all \( i \) large, \( v + \left( {\delta /2}\right) {u}_{i} \in {T}_{S}^{C}\left( {x}_{i}\right) \), whence \[ \left\langle {{\zeta }_{i}, v + \left( \frac{\delta }{2}\right) {u}_{i}}\right\rangle \leq 0 \] This implies \( \left\langle {{\zeta }_{i}, v}\right\rangle \leq \left( {\delta /2}\right) \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \), and in the limit \( \langle \zeta, v\rangle \leq \delta /2 \), the desired contradiction. The sufficiency part of the proof is no harder than the necessity; it constitutes one of the end-of-chapter problems. A cone \( K \) in \( {\mathbb{R}}^{n} \) is called pointed if it contains no two nonzero elements whose sum is zero. 6.9. Exercise. A convex cone \( K \) in \( {\mathbb{R}}^{n} \) has nonempty interior iff its polar \( {K}^{ \circ } \) is pointed. We can summarize some previous results in terms of pointedness of the normal cone, as
1012_(GTM178)Nonsmooth Analysis and Control Theory
60
delta /2}\right) B \) for all \( i \) sufficiently large. Then, for some vector \( {u}_{i} \in B \), and for all \( i \) large, \( v + \left( {\delta /2}\right) {u}_{i} \in {T}_{S}^{C}\left( {x}_{i}\right) \), whence \[ \left\langle {{\zeta }_{i}, v + \left( \frac{\delta }{2}\right) {u}_{i}}\right\rangle \leq 0 \] This implies \( \left\langle {{\zeta }_{i}, v}\right\rangle \leq \left( {\delta /2}\right) \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \), and in the limit \( \langle \zeta, v\rangle \leq \delta /2 \), the desired contradiction. The sufficiency part of the proof is no harder than the necessity; it constitutes one of the end-of-chapter problems. A cone \( K \) in \( {\mathbb{R}}^{n} \) is called pointed if it contains no two nonzero elements whose sum is zero. 6.9. Exercise. A convex cone \( K \) in \( {\mathbb{R}}^{n} \) has nonempty interior iff its polar \( {K}^{ \circ } \) is pointed. We can summarize some previous results in terms of pointedness of the normal cone, as follows: 6.10. Corollary. If \( {N}_{S}^{C}\left( x\right) \) is pointed, then \( {N}_{S}^{C}\left( \cdot \right) \) is graph-closed at \( x \) , \( {T}_{S}^{C}\left( \cdot \right) \) is lower semicontinuous at \( x \), and \( S \) is wedged at \( x \) . ## 6.11. Exercise. (a) When \( n = 1,{N}_{S}^{C}\left( \cdot \right) \) is always graph-closed at each point of \( S \) . (b) We construct an example in \( {\mathbb{R}}^{2} \) of a set \( S \) such that \( {N}_{S}^{C}\left( \cdot \right) \) fails to be graph-closed at the origin; \( S \) will be the graph of a certain continuous function \( f : \left\lbrack {0,1}\right\rbrack \rightarrow \mathbb{R} \) . We set \( f\left( x\right) = 0 \) at every point \( x \) of the form \( {2}^{-n} \) for \( n = 0,1,2,\ldots \) and we set \( f\left( 0\right) = 0 \) . In between any two points of the form \( {2}^{-n - 1} \) and \( {2}^{-n} \), the graph of \( f \) describes an isosceles triangle whose apex is located at \( \left( {\left( {{2}^{-n - 1} + {2}^{-n}}\right) /2,{2}^{-{2n}}}\right) \) . Using the Proximal Normal Formula of Theorem 2.6.1 for \( {N}_{S}^{C} \), show that \( {N}_{S}^{C}\left( {0,0}\right) \) is a half-space. Show also that \( {N}_{S}^{C} = {\mathbb{R}}^{2} \) at each point of the form \( \left( {{2}^{-n},0}\right), n > 1 \) . Conclude that \( {N}_{S}^{C} \) is not graph-closed at \( \left( {0,0}\right) \) . ## A General Relation Between \( {T}_{S}^{C} \) and \( {T}_{S}^{B} \) We now establish a result which clarifies the relationship between the two notions of tangency. It affirms that asymptotically, a vector \( v \) which lies in \( {T}_{S}^{C}\left( x\right) \) is one which lies in \( {T}_{S}^{B}\left( {x}^{\prime }\right) \) for all \( {x}^{\prime } \) near \( x \) . 6.12. Theorem. \( v \in {T}_{S}^{C}\left( x\right) \) iff \[ \mathop{\limsup }\limits_{{{x}^{\prime }\overset{S}{ \rightarrow }x}}d\left( {v,{T}_{S}^{B}\left( {x}^{\prime }\right) }\right) = 0. \] Proof. To begin with, let \( v \) satisfy the given limit condition, and let us show that \( v \in {T}_{S}^{C}\left( x\right) \) . It suffices to show that \( \langle v,\zeta \rangle \leq 0 \) for any element \( \zeta \in {N}_{S}^{L}\left( x\right) \), since \( {N}_{S}^{L}{\left( x\right) }^{ \circ } = {N}_{S}^{C}{\left( x\right) }^{ \circ } = {T}_{S}^{C}\left( x\right) \) . Such a vector \( \zeta \) is of the form \[ \mathop{\lim }\limits_{{i \rightarrow \infty }}{\zeta }_{i},\;{\zeta }_{i} \in {N}_{S}^{P}\left( {x}_{i}\right) ,\;{x}_{i}\overset{S}{ \rightarrow }x. \] By hypothesis, there exists \( {v}_{i} \in {T}_{S}^{B}\left( {x}_{i}\right) \) such that \( {v}_{i} \rightarrow v \) . Recall now that \( {T}_{S}^{B}\left( {x}_{i}\right) \subset {N}_{S}^{P}{\left( {x}_{i}\right) }^{ \circ } \) (Exercise 2.7.1); in consequence we have \( \left\langle {{v}_{i},{\zeta }_{i}}\right\rangle \leq 0 \) . In the limit we get \( \langle \zeta, v\rangle \leq 0 \), as required. Suppose now that \( v \in {T}_{S}^{C}\left( x\right) \), but that the limit condition of the theorem fails. Then, for some \( \varepsilon > 0 \) and sequence \( {x}_{i} \rightarrow x,{x}_{i} \in S \), we have \( d\left( {v,{T}_{S}^{B}\left( {x}_{i}\right) }\right) > \varepsilon \) . Since a vector \( u \) lies in \( {T}_{S}^{B}\left( y\right) \) iff \( D{d}_{S}\left( {y;u}\right) \leq 0 \) (by Exercise 2.7.1) we have \( D{d}_{S}\left( {{x}_{i};v + w}\right) > 0 \) for all \( w \in \varepsilon \bar{B} \) . Subbotin’s Theorem 4.2 applies to give \( {z}_{i} \in {x}_{i} + \left( {1/i}\right) B \) admitting \( {\zeta }_{i} \in {\partial }_{P}{d}_{S}\left( {z}_{i}\right) \) such that \[ \left\langle {{\zeta }_{i}, v + w}\right\rangle > 0\forall w \in \varepsilon \bar{B} \] We deduce \( \left\langle {{\zeta }_{i}/\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix}, v}\right\rangle \geq \varepsilon \) . Taking a subsequence if necessary to arrange the convergence of \( {\zeta }_{i}/\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \) to a limit \( \zeta \), we have \( \zeta \in {N}_{S}^{L}\left( x\right) \) (since \( {\zeta }_{i} \in \) \( \left. {{\partial }_{P}{d}_{S}\left( {z}_{i}\right) \subset {N}_{S}^{P}\left( {z}_{i}\right) }\right) \) and \( \langle \zeta, v\rangle \geq \varepsilon \) . This contradicts \( v \in {N}_{S}^{L}{\left( x\right) }^{ \circ } = {T}_{S}^{C}\left( x\right) \) , and completes the proof. Recall that \( S \) is said to be regular at \( x \) if \( {T}_{S}^{B}\left( x\right) = {T}_{S}^{C}\left( x\right) \) . 6.13. Corollary. \( S \) is regular at \( x \) iff \( {T}_{S}^{B}\left( \cdot \right) \) is lower semicontinuous at \( x \) . 6.14. Exercise. (a) Prove Corollary 6.12. (b) Show that the set \( S \) of Exercise \( {6.11}\left( \mathrm{\;b}\right) \) is regular at \( \left( {0,0}\right) \), but that \( {T}_{S}^{C}\left( \cdot \right) \) is not lower semicontinuous there. (c) A continuous function \( f \) is such that \( f\left( x\right) \in {T}_{S}^{B}\left( x\right) \) for every \( x \) iff it satisfies \( f\left( x\right) \in {T}_{S}^{C}\left( x\right) \) for every \( x \) . ## 7 Problems on Chapter 3 7.1. Let \( f \) and \( g \) be locally Lipschitz functions on \( {\mathbb{R}}^{n} \) satisfying \( f \geq g \) . Suppose that at a point \( {x}_{0} \) we have \( f\left( {x}_{0}\right) = g\left( {x}_{0}\right) \) . Prove that \[ {\partial }_{C}f\left( {x}_{0}\right) \cap {\partial }_{C}g\left( {x}_{0}\right) \neq \varnothing . \] Show that the corresponding result is false if \( {\partial }_{C} \) is replaced by either \( {\partial }_{L} \) or \( {\partial }_{D} \) . 7.2. We adopt the notation and the hypotheses of Theorem 1.7. (a) Obtain the following "inner estimate" for \( {\partial }_{C}V\left( 0\right) \) : For any \( x \in \sum \left( 0\right) \) , we have \( {\partial }_{C}V\left( 0\right) \cap M\left( x\right) \neq \varnothing \) . (Hint. Consider the device used in Exercise 1.5, in light of the preceding problem.) (b) We have \[ {DV}\left( {0;u}\right) \geq \mathop{\inf }\limits_{{x \in \sum \left( 0\right) }}\mathop{\inf }\limits_{{\zeta \in M\left( x\right) }}\langle \zeta, u\rangle \] \[ D\left( {-V}\right) \left( {0;u}\right) \geq \mathop{\sup }\limits_{{x \in \sum \left( 0\right) }}\mathop{\inf }\limits_{{\zeta \in M\left( x\right) }}\langle - \zeta, u\rangle . \] (c) If \( M\left( x\right) \) is a singleton \( \{ \zeta \left( x\right) \} \) for each \( x \in \sum \left( 0\right) \), then \( {V}^{\prime }\left( {0;u}\right) \) exists for each \( u \) and equals \( \mathop{\inf }\limits_{{x \in \sum \left( 0\right) }}\langle \zeta \left( x\right), u\rangle \) . (d) If \( \sum \left( 0\right) \) is a singleton \( \{ x\} \) and \( M\left( x\right) \) is a singleton \( \{ \zeta \} \), then \( {V}^{\prime }\left( 0\right) \) exists, and we have \( \nabla V\left( 0\right) = \zeta \) . 7.3. A function \( f \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) is called calm at \( x \in \operatorname{dom}f \) if it satisfies \[ \mathop{\liminf }\limits_{{{x}^{\prime } \rightarrow x}}\frac{f\left( {x}^{\prime }\right) - f\left( x\right) }{\begin{Vmatrix}{x}^{\prime } - x\end{Vmatrix}} > - \infty . \] Prove that if \( f \) is calm at \( x \), then for some \( K > 0 \) and for any \( \varepsilon > 0 \), there exist \( z \in x + {\varepsilon B} \) and \( \zeta \in {\partial }_{P}f\left( z\right) \) with \( \left| {f\left( z\right) - f\left( x\right) }\right| < \varepsilon \) and \( \parallel \zeta \parallel \leq K \) (i.e., \( f \) admits a priori bounded proximal subgradients arbitrarily near \( x \) ). 7.4. Prove the following addendum to Exercise 1.8. If \( V\left( \cdot \right) \) is calm at 0, then we may take \( {\lambda }_{0} = 1 \) . 7.5. We adopt the notation and hypotheses of Theorem 1.14, and we assume in addition that the functions \( f,{g}_{i}\left( {i = 1,2,\ldots, p}\right) \) are convex, and that the functions \( {h}_{j}\left( {j = 1,2,\ldots, m}\right) \) are affine. (a) Prove that \( V \) is convex. (b) Prove that if every \( x \in \sum \left( {0,0}\right) \) is normal, then \[ {\partial }_{L}V\left( {0,0}\right) = {\partial }_{C}V\left( {0,0}\right) = \mathop{\bigcup }\limits_{{x \in \sum \left( {0,0}\right) }}M\left( x\right) . \] (c) Prove that any \( x \) feasible for \( P\left( {0,0}\right) \) for which \( M\left( x\right) \neq \varnothing \) is a solution to \( P\left( {0,0}\right) \) . (This is a general principle in optimization: In the convex case of the problem, the (normal) necessary conditions become sufficient as well.) 7.6. Consider the smooth, finite-dimensional case of the Mean Value Inequality, Theorem 2.3, when \( n = 2, x = \left( {0,0}\right) \) , \[ Y = \{ \left( {1, t}\right) : 0 \leq t \leq 1\} ,\;f\left( {u, v}\right) = u + \left( {1 - u}\right) {v}^{2}. \] Find the points \( z \) satisfying the conclusion of the theorem. Observe that none of them lie in int \( \left\lbrack {0, Y}\right\rbrack \) . 7.7. Consider the general case of the Mean Value Inequality, Theorem 2.
1012_(GTM178)Nonsmooth Analysis and Control Theory
61
every \( x \in \sum \left( {0,0}\right) \) is normal, then \[ {\partial }_{L}V\left( {0,0}\right) = {\partial }_{C}V\left( {0,0}\right) = \mathop{\bigcup }\limits_{{x \in \sum \left( {0,0}\right) }}M\left( x\right) . \] (c) Prove that any \( x \) feasible for \( P\left( {0,0}\right) \) for which \( M\left( x\right) \neq \varnothing \) is a solution to \( P\left( {0,0}\right) \) . (This is a general principle in optimization: In the convex case of the problem, the (normal) necessary conditions become sufficient as well.) 7.6. Consider the smooth, finite-dimensional case of the Mean Value Inequality, Theorem 2.3, when \( n = 2, x = \left( {0,0}\right) \) , \[ Y = \{ \left( {1, t}\right) : 0 \leq t \leq 1\} ,\;f\left( {u, v}\right) = u + \left( {1 - u}\right) {v}^{2}. \] Find the points \( z \) satisfying the conclusion of the theorem. Observe that none of them lie in int \( \left\lbrack {0, Y}\right\rbrack \) . 7.7. Consider the general case of the Mean Value Inequality, Theorem 2.6, when \( X = \mathbb{R}, x = 0, Y = \{ 1\} \), and \[ f\left( u\right) = \left\{ \begin{array}{ll} - \sqrt{\left| u\right| } & \text{ if }u \leq 0 \\ 1 & \text{ if }u > 0 \end{array}\right. \] Show that for all \( \bar{r} \) and \( \varepsilon \) sufficiently near \( \widehat{r} \) and 0, respectively, the point \( z \) whose existence is asserted by the theorem must lie outside \( \left\lbrack {x, Y}\right\rbrack \) . 7.8. Let \( S \) and \( E \) be two bounded, closed, nonempty subsets of the Hilbert space \( X \), with \( 0 \notin E \) . Prove that for some \( s \in S \) and \( \zeta \in {N}_{S}^{P}\left( s\right) \), we have \( \langle \zeta, e\rangle > 0 \) for all \( e \in E \) . 7.9. Show that the Decrease Principle, Theorem 2.8, remains true under the following weakened hypothesis: For some \( \Delta > 0,\rho > 0 \), and \( \delta > 0 \), we have \[ z \in {x}_{0} + {\rho B},\;\zeta \in {\partial }_{P}f\left( z\right) ,\;f\left( z\right) < f\left( {x}_{0}\right) + \Delta \Rightarrow \parallel \zeta \parallel \geq \delta . \] 7.10. Let \( f : {\mathbb{R}}^{2} \rightarrow \mathbb{R} \) be differentiable, with \( f\left( {-1, v}\right) < 0 \) and \( f\left( {1, v}\right) > 0 \) for all \( v \in \left\lbrack {-1,1}\right\rbrack \) . Prove the existence of a point \( \left( {x, y}\right) \) satisfying \[ \left| y\right| \leq \left| x\right| \leq 1\text{ and }{f}_{x}\left( {x, y}\right) > \left| {{f}_{y}\left( {x, y}\right) }\right| . \] 7.11. Show that for some \( \delta > 0 \), for all \( \left( {u, v}\right) \) in \( {\mathbb{R}}^{2} \) sufficiently near \( \left( {0,0}\right) \) , there exists a solution \( \left( {x, y, z}\right) \) in \( {\mathbb{R}}^{3} \) to the system \[ \left| {x - y}\right| + {2z} = u,\;x - \left| z\right| = v, \] satisfying \[ \parallel \left( {x, y, z}\right) \parallel \leq \frac{1}{\delta }\parallel \left( {u, v}\right) \parallel \] 7.12. Show that for some \( \delta > 0 \), for all \( \left( {x, y, z}\right) \) in \( {\mathbb{R}}^{3} \) sufficiently near \( \left( {0,0,0}\right) \), there exists \( \left( {{x}^{\prime },{y}^{\prime },{z}^{\prime }}\right) \) satisfying \[ \left| {{x}^{\prime } - {y}^{\prime }}\right| + 2{z}^{\prime } = 0,\;{x}^{\prime } - \left| {z}^{\prime }\right| = 0, \] such that \[ \begin{Vmatrix}{\left( {{x}^{\prime },{y}^{\prime },{z}^{\prime }}\right) - \left( {x, y, z}\right) }\end{Vmatrix} \leq \frac{1}{\delta }\parallel \left( {\left| {x - y}\right| + {2z}, x - \left| z\right| }\right) \parallel . \] 7.13. (a) Let \( G : {\mathbb{R}}^{m} \rightarrow {\mathbb{R}}^{n} \) be Lipschitz near \( x \), and let \( {\partial }_{C}G\left( x\right) \) be of maximal rank. Then we have \[ 0 \in {\partial }_{L}\langle \zeta, G\left( \cdot \right) \rangle \left( x\right) \Rightarrow \zeta = 0. \] (b) Verify that \( {\partial }_{C}G\left( 0\right) \) is of maximal rank when \( m = 3, n = 2 \), and \[ G\left( {x, y, z}\right) \mathrel{\text{:=}} \left( {\left| {x - y}\right| + {2z}, x - \left| z\right| }\right) . \] (What is the relevance of this to the two preceding problems?) 7.14. Let \( {x}_{0} \in {S}_{1} \cap {S}_{2} \), where \( {S}_{1} \) and \( {S}_{2} \) are closed subsets of \( {\mathbb{R}}^{n} \) satisfying \[ {N}_{{S}_{1}}^{L}\left( {x}_{0}\right) \cap \left( {-{N}_{{S}_{2}}^{L}\left( {x}_{0}\right) }\right) = \{ 0\} . \] Prove that for some \( \delta > 0 \), for all \( x \) sufficiently near \( {x}_{0} \), we have \[ {d}_{{S}_{1} \cap {S}_{2}}\left( x\right) \leq \frac{1}{\delta }\max \left\{ {{d}_{{S}_{1}}\left( x\right) ,{d}_{{S}_{2}}\left( x\right) }\right\} . \] 7.15. Let the functions \( g, h \) be as in Theorem 3.8, and in particular continue to assume the constaint qualification at \( {x}_{0} \) . Let \( S \) be the set \[ \{ x \in X : g\left( x\right) \leq 0, h\left( x\right) = 0\} . \] We assume that \( g\left( {x}_{0}\right) = 0 \), and that \( h \) is \( {C}^{1} \) near \( {x}_{0} \) . We set \[ I \mathrel{\text{:=}} \{ 1,2,\ldots, p\} ,\;J \mathrel{\text{:=}} \{ 1,2,\ldots, m\} . \] (a) \( {T}_{S}^{C}\left( {x}_{0}\right) \) contains every vector \( v \in X \) satisfying \[ {g}_{i}^{ \circ }\left( {{x}_{0};v}\right) \leq 0\;\left( {i \in I}\right) ,\;\left\langle {{h}_{j}^{\prime }\left( {x}_{0}\right), v}\right\rangle = 0\;\left( {j \in J}\right) . \] (b) If each of the functions \( {g}_{i} \) admits directional derivatives at \( {x}_{0} \), then any vector \( v \in {T}_{S}^{B}\left( {x}_{0}\right) \) satisfies \[ {g}_{i}^{\prime }\left( {{x}_{0};v}\right) \leq 0\;\left( {i \in I}\right) ,\;{h}_{j}^{\prime }\left( {{x}_{0};v}\right) = 0\;\left( {j \in J}\right) . \] (c) If each of the functions \( {g}_{i} \) is regular at \( {x}_{0} \), then \( S \) is regular at \( {x}_{0} \) , and we have \[ {T}_{S}^{C}\left( {x}_{0}\right) = {T}_{S}^{B}\left( {x}_{0}\right) \] \[ = \left\{ {v : {g}_{i}^{\prime }\left( {{x}_{0};v}\right) \leq 0\left( {i \in I}\right) ,\left\langle {{h}_{j}^{\prime }\left( {x}_{0}\right), v}\right\rangle = 0\left( {j \in J}\right) }\right\} , \] \[ {N}_{S}^{C}\left( {x}_{0}\right) = \operatorname{cone}\left\{ {\zeta \in {\partial }_{C}\{ \langle \gamma, g\left( \cdot \right) \rangle +\langle \lambda, h\left( \cdot \right) \rangle \} \left( {x}_{0}\right) }\right. \] \[ \left. { : \gamma \in {\mathbb{R}}^{p},\lambda \in {\mathbb{R}}^{m},\gamma \geq 0}\right\} \text{.} \] (Hint. Part (a) can be proven via the characterization provided by Proposition 2.5.2, together with an appeal to Theorem 3.8.) 7.16. (a) Let \( f : {\mathbb{R}}^{2} \rightarrow \mathbb{R} \) be defined as follows: \[ f\left( {x, y}\right) = \left\{ \begin{array}{ll} - {\left| x\right| }^{3/2} & \text{ if }x \leq 0 \\ 0 & \text{ if }x \geq 0 \\ \min \left( {x, y}\right) & \text{ if }x \geq 0 \\ \min \left( {x, y}\right) & \text{ if }x \geq 0 \end{array}\right. \] Prove that \( f \) is Lipschitz near \( \left( {0,0}\right) \) . Calculate the sets \( {\partial }_{P}f\left( {0,0}\right) \) , \( {\partial }_{D}f\left( {0,0}\right) ,{\partial }_{L}f\left( {0,0}\right) \), and \( {\partial }_{C}f\left( {0,0}\right) \), and note that they are all different. (b) Calculate \( {\partial }_{D}f\left( 0\right) \) where \( f \) is the function that appears in Problem 2.9.16. 7.17. Let \( x \in S \), where \( S \) is a closed subset of \( {\mathbb{R}}^{n} \) . (a) Prove that \( {\partial }_{D}{d}_{S}\left( x\right) \subset {N}_{S}^{D}\left( x\right) \cap \bar{B} \) . (b) Prove now that equality holds. (Hint. Use Exercise 4.9(c) and Proposition 4.10.) (c) Deduce from (b) the corresponding fact for limiting constructs: \[ {\partial }_{L}{d}_{S}\left( x\right) = {N}_{S}^{L}\left( x\right) \cap \bar{B} \] (d) Show that \( {\partial }_{C}{d}_{S}\left( x\right) \) and \( {N}_{S}^{C}\left( x\right) \cap \bar{B} \) differ in general. 7.18. Let \( x \in S \), where \( S \) is a closed subset of \( {\mathbb{R}}^{n} \) . We will prove that \( S \) is regular at \( x \) iff \( {d}_{S} \) is regular at \( x \) . (a) Let \( S \) be regular at \( x \) . Use the preceding exercise to justify the following steps: \[ {d}_{S}^{ \circ }\left( {x;v}\right) = \max \left\{ {\langle \zeta, v\rangle : \zeta \in {\partial }_{L}{d}_{S}\left( x\right) }\right\} \] \[ = \max \left\{ {\langle \zeta, v\rangle : \zeta \in {N}_{S}^{L}\left( x\right) \cap \bar{B}}\right\} \] \[ = \max \left\{ {\langle \zeta, v\rangle : \zeta \in {N}_{S}^{D}\left( x\right) \cap \bar{B}}\right\} \] \[ = \max \left\{ {\langle \zeta, v\rangle : \zeta \in {\partial }_{D}{d}_{S}\left( x\right) }\right\} \] \[ \leq D{d}_{S}\left( {x;v}\right) \text{.} \] Deduce that \( {d}_{S} \) is regular at \( x \) . (b) If \( {d}_{S}\left( \cdot \right) \) is regular at \( x \), show that \( S \) is regular at \( x \) . 7.19. Let \( E \) be a compact subset of \( {\mathbb{R}}^{n} \), and let \( S \) consist of those functions \( x\left( \cdot \right) \in {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) such that \( \dot{x}\left( t\right) \in E \) a.e. Prove that \( S \) is weakly compact iff \( E \) is convex. What is the weak closure of \( S \) if \( E \) is not convex? 7.20. Let \( \varphi : {\mathbb{R}}^{n} \times {\mathbb{R}}^{m} \rightarrow {\mathbb{R}}^{k} \) be such that for every \( x \in {\mathbb{R}}^{n} \), the function \( u \mapsto \varphi \left( {x, u}\right) \) is continuous, and for every \( u \in {\mathbb{R}}^{m} \), the function \( x \mapsto \varphi \left( {x, u}\right) \) is measurable. Suppose that \( \Gamma \) is a closed-valued measurable multifunction from \( {\mathbb{R}}^{n} \) to \( {\mathbb{R}}^{m} \) such that \[ \{ u \in \Gamma \left( x\right) : \varphi \left( {x, u}\right) = 0\} \neq \varnothing \forall x \in {\mathbb{R}}^{n}. \] Prove the existence of a measurable selection \( \gamma \) for \( \Gamma \) such that \[ \varphi \left( {x,\gamma \left( x\right) }\right) = 0\forall x \in {\mathbb{R}}^{n}. \] (This is known as Filippov's Lemma.) 7.21. Let \( \Gamma \) be
1012_(GTM178)Nonsmooth Analysis and Control Theory
62
dot{x}\left( t\right) \in E \) a.e. Prove that \( S \) is weakly compact iff \( E \) is convex. What is the weak closure of \( S \) if \( E \) is not convex? 7.20. Let \( \varphi : {\mathbb{R}}^{n} \times {\mathbb{R}}^{m} \rightarrow {\mathbb{R}}^{k} \) be such that for every \( x \in {\mathbb{R}}^{n} \), the function \( u \mapsto \varphi \left( {x, u}\right) \) is continuous, and for every \( u \in {\mathbb{R}}^{m} \), the function \( x \mapsto \varphi \left( {x, u}\right) \) is measurable. Suppose that \( \Gamma \) is a closed-valued measurable multifunction from \( {\mathbb{R}}^{n} \) to \( {\mathbb{R}}^{m} \) such that \[ \{ u \in \Gamma \left( x\right) : \varphi \left( {x, u}\right) = 0\} \neq \varnothing \forall x \in {\mathbb{R}}^{n}. \] Prove the existence of a measurable selection \( \gamma \) for \( \Gamma \) such that \[ \varphi \left( {x,\gamma \left( x\right) }\right) = 0\forall x \in {\mathbb{R}}^{n}. \] (This is known as Filippov's Lemma.) 7.21. Let \( \Gamma \) be a multifunction from \( {\mathbb{R}}^{m} \) to \( {\mathbb{R}}^{n} \) whose images are compact and convex. Prove that \( \Gamma \) is measurable iff the function \( H\left( {x, p}\right) \) is measurable in \( x \) for each \( p \), where \[ H\left( {x, p}\right) \mathrel{\text{:=}} \sup \{ \langle p, v\rangle : v \in \Gamma \left( x\right) \} . \] 7.22. Consider \( f : {L}_{1}^{2}\left\lbrack {0,1}\right\rbrack \rightarrow \mathbb{R} \) defined by \[ f\left( x\right) \mathrel{\text{:=}} {\int }_{0}^{1}\varphi \left( {x\left( t\right) }\right) {dt},\text{ with }\varphi \left( x\right) \mathrel{\text{:=}} - \left| x\right| . \] (a) Prove that \( {\partial }_{P}f\left( x\right) \neq \varnothing \) iff there exists \( \delta > 0 \) such that \[ \operatorname{meas}\{ t \in \left\lbrack {0,1}\right\rbrack : - \delta \leq x\left( t\right) \leq \delta \} = 0. \] (b) Exhibit \( \zeta \) and \( x \in {L}_{1}^{2}\left\lbrack {0,1}\right\rbrack \) such that \( \zeta \left( t\right) \in {\partial }_{P}\varphi \left( {x\left( t\right) }\right) \) a.e., yet \( \zeta \notin \) \( {\partial }_{P}f\left( x\right) \) . 7.23. Show that any absolutely continuous function \( x \) satisfying the Euler inclusion and transversality condition of Theorem 5.22 is a solution of the variational problem when \( \varphi \) and \( \ell \) are convex. 7.24. In the context of Theorem 5.22, let \( \varphi \left( {u, v}\right) \) have the form \( (1 + \) \( \parallel v{\parallel }^{2}{)}^{1/2} + g\left( u\right) \), where \( g \) is Lipschitz. Prove that any solution \( x \) of the variational problem is continuously differentiable. 7.25. Prove the sufficiency part of Proposition 6.8: If \( {N}_{S}^{C}\left( \cdot \right) \) is graph-closed at \( x \), then \( {T}_{S}^{C}\left( \cdot \right) \) is lower semicontinuous at \( x \) . 7.26. Let \( S \) be a closed convex subset of \( {\mathbb{R}}^{n} \), and let \( x \in S \) . Then \( S \) is wedged at \( x \) iff int \( S \neq \varnothing \) . 7.27. Let \( S \) be a closed subset of \( {\mathbb{R}}^{n} \) which is wedged at \( x \), where \( x \in \operatorname{bdry}S \) . Then \[ {N}_{S}^{C}\left( x\right) = - {N}_{\widehat{S}}^{C}\left( x\right) \] where \( \widehat{S} \) is the set \( \operatorname{cl}\left( {\operatorname{comp}S}\right) \) . Find a counterexample to this formula when \( S \) is not wedged at \( x \) . 7.28. Let \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be locally Lipschitz. Prove that epi \( f \) is wedged at each of its points. (Conversely, a set which is wedged can be viewed locally as the epigraph of a Lipschitz function, for a suitable choice of coordinate axes, as can be shown.) 7.29. The Hausdorff distance \( \rho \left( {C, S}\right) \) between two sets \( C \) and \( S \) in \( {\mathbb{R}}^{n} \) is defined via \[ \rho \left( {C, S}\right) = \max \left\{ {\mathop{\sup }\limits_{{c \in C}}{d}_{S}\left( c\right) ,\mathop{\sup }\limits_{{s \in S}}{d}_{C}\left( s\right) }\right\} . \] A multifunction \( \Gamma \) from \( {\mathbb{R}}^{m} \) to \( {\mathbb{R}}^{n} \) is (Hausdorff) continuous at \( x \) if \[ \rho \left( {\Gamma \left( {x}^{\prime }\right) ,\Gamma \left( x\right) }\right) \rightarrow 0\text{ as }{x}^{\prime } \rightarrow x. \] Prove that when \( \Gamma \) is closed-valued, then the continuity of \( \Gamma \) at \( x \) implies that \( \Gamma \) is both upper and lower semicontinuous at \( x \) . 7.30. It is known that there exists a subset \( S \) of \( \mathbb{R} \) with the following property: for every \( a, b \in \mathbb{R} \) with \( a < b \), the Lebesgue measure \( \mathcal{L} \) of the set \( S \cap \left\lbrack {a, b}\right\rbrack \) satisfies \[ \frac{\left( b - a\right) }{3} < \mathcal{L}\left( {S \cap \left\lbrack {a, b}\right\rbrack }\right) < \frac{2\left( {b - a}\right) }{3}. \] Let \( f : \mathbb{R} \rightarrow \mathbb{R} \) be defined as follows: \[ f\left( x\right) = {\int }_{0}^{x}{\chi }_{S}\left( t\right) {dt} \] where \( {\chi }_{S} \) is the characteristic function of \( S \) . Prove the following facts: (a) \( f \) is Lipschitz, and we have \( {\partial }_{C}f\left( x\right) = \left\lbrack {0,1}\right\rbrack \) for all \( x \in \mathbb{R} \) . (b) The functions \( f\left( x\right) \) and \( g\left( x\right) \mathrel{\text{:=}} x - f\left( x\right) \) have the same generalized gradient at every point, yet do not differ by a constant. We remark that \( {\partial }_{P}f \) and \( {\partial }_{L}f \) are not known, nor is it known whether the phenomenon that occurs in (b) can take place in terms of these subdiffer-entials. 4 A Short Course in Control Theory We are guided by the beauty of our weapons. Mathematics, as well as several areas of application, abounds with situations where it is desired to control the behavior of the trajectories of a given dynamical system. The goal can be either geometric (keep the state of the system in a given set, or bring it toward the set), or functional (find the trajectory that is optimal relative to a given criterion). More specific issues arise subsequently, such as the construction of feedback control mechanisms achieving the aims we have in mind. In this chapter we will identify a complex of such fundamental and related issues, as they arise in connection with the control of ordinary differential equations in a deterministic setting. The first section sets the scene and develops a technical base for the entire chapter. ## 1 Trajectories of Differential Inclusions We are given a multifunction \( F \) mapping \( \mathbb{R} \times {\mathbb{R}}^{n} \) to the subsets of \( {\mathbb{R}}^{n} \), and a time interval \( \left\lbrack {a, b}\right\rbrack \) . The central object of study in this chapter will be the differential inclusion \[ \dot{x}\left( t\right) \in F\left( {t, x\left( t\right) }\right) \text{ a.e.,}\;t \in \left\lbrack {a, b}\right\rbrack . \] (1) A solution \( x\left( \cdot \right) \) of (1) is taken to mean an absolutely continuous function \( x : \left\lbrack {a, b}\right\rbrack \rightarrow {\mathbb{R}}^{n} \) which, together with \( \dot{x} \), its derivative with respect to \( t \) , satisfies (1). For brevity, we will refer to any absolutely continuous \( x \) from \( \left\lbrack {a, b}\right\rbrack \) to \( {\mathbb{R}}^{n} \) as an \( {arc} \) on \( \left\lbrack {a, b}\right\rbrack \) . We also refer to an arc \( x \) satisfying (1) as a trajectory of \( F \) . The concept of differential inclusion subsumes that of a standard control system \[ \dot{x}\left( t\right) = f\left( {t, x\left( t\right), u\left( t\right) }\right) \] (2) where \( f : \mathbb{R} \times {\mathbb{R}}^{n} \times {\mathbb{R}}^{m} \rightarrow {\mathbb{R}}^{n} \), and where the control function \( u \) takes values in some prescribed subset \( U \) of \( {\mathbb{R}}^{m} \) ; simply consider \( F\left( {t, x}\right) \mathrel{\text{:=}} f\left( {t, x, U}\right) \) . Filippov's Lemma (Problem 3.7.20) implies that under mild hypotheses on \( f \), an arc \( x \) satisfies (1) iff there is a measurable function \( u\left( \cdot \right) \) with values in \( U \) such that (2) holds. A more special case of (1) is the one in which \( F\left( {t, x}\right) \) is a singleton for all \( \left( {t, x}\right) \) ; i.e., \( F\left( {t, x}\right) = \{ f\left( {t, x}\right) \} \) . In the classical study of the ordinary differential equation \[ \dot{x}\left( t\right) = f\left( {t, x\left( t\right) }\right) \] (3) the properties of the function \( f \) play an important role. Let us recall some basic facts on this subject: 1.1. Theorem. Suppose that \( f \) is continuous, and let \( \left( {{t}_{0},{x}_{0}}\right) \in \mathbb{R} \times {\mathbb{R}}^{n} \) be given. Then the following hold: (a) There exists a solution of (3) on an open interval \( \left( {{t}_{0} - \delta ,{t}_{0} + \delta }\right) \), for some \( \delta > 0 \), satisfying \( x\left( {t}_{0}\right) = {x}_{0} \) . (b) If in addition we assume linear growth, that is that there exist nonnegative constants \( \gamma \) and \( c \) such that \[ \parallel f\left( {t, x}\right) \parallel \leq \gamma \parallel x\parallel + c\forall \left( {t, x}\right) , \] then there exists a solution of (3) on \( \left( {-\infty ,\infty }\right) \) such that \( x\left( {t}_{0}\right) = {x}_{0} \) . (c) Let us now add the hypothesis that \( f \) is locally Lipschitz. Then there exists a unique solution of (3) on \( \left( {-\infty ,\infty }\right) \) such that \( x\left( {t}_{0}\right) = {x}_{0} \) . Many readers will know that the hypotheses on \( f \) in the preceding theorem can be relaxed - for example, joint continuity of \( f \) in \( \left( {t, x}\right) \) can be replaced with measurability in \( t \) and continuity in \( x \) . However, the form given above will suffice for our purposes. In developing the basic theory of differential inclusions, two properties of \( F \) turn out to be particularly important: upper semicontinuity and
1012_(GTM178)Nonsmooth Analysis and Control Theory
63
inear growth, that is that there exist nonnegative constants \( \gamma \) and \( c \) such that \[ \parallel f\left( {t, x}\right) \parallel \leq \gamma \parallel x\parallel + c\forall \left( {t, x}\right) , \] then there exists a solution of (3) on \( \left( {-\infty ,\infty }\right) \) such that \( x\left( {t}_{0}\right) = {x}_{0} \) . (c) Let us now add the hypothesis that \( f \) is locally Lipschitz. Then there exists a unique solution of (3) on \( \left( {-\infty ,\infty }\right) \) such that \( x\left( {t}_{0}\right) = {x}_{0} \) . Many readers will know that the hypotheses on \( f \) in the preceding theorem can be relaxed - for example, joint continuity of \( f \) in \( \left( {t, x}\right) \) can be replaced with measurability in \( t \) and continuity in \( x \) . However, the form given above will suffice for our purposes. In developing the basic theory of differential inclusions, two properties of \( F \) turn out to be particularly important: upper semicontinuity and the Lipschitz condition. We will not see the latter intervene until the next section, but the former property is part of the Standing Hypotheses on \( F \) that will be in force throughout the rest of this chapter, whether explicitly mentioned or not. ## 1.2. Standing Hypotheses. (a) For every \( \left( {t, x}\right), F\left( {t, x}\right) \) is a nonempty compact convex set. (b) \( F \) is upper semicontinuous. (c) For some positive constants \( \gamma \) and \( c \), and for all \( \left( {t, x}\right) \) , \[ v \in F\left( {t, x}\right) \Rightarrow \parallel v\parallel \leq \gamma \parallel x\parallel + c. \] Parts (b) and (c) of the Standing Hypotheses are of course analogues of the continuity and linear growth conditions, respectively, familiar in the classical case. We recall that \( F \) is upper semicontinuous at \( x \) if, given any \( \varepsilon > 0 \), there exists \( \delta > 0 \) such that \[ \begin{Vmatrix}{{x}^{\prime } - x}\end{Vmatrix} < \delta \Rightarrow F\left( {x}^{\prime }\right) \subset F\left( x\right) + {\varepsilon B}. \] 1.3. Exercise. (a) When \( F\left( {t, x}\right) = \{ f\left( {t, x}\right) \} \), show that Standing Hypotheses 1.2 (b, c) hold iff \( f \) is continuous and satisfies \( \parallel f\left( {t, x}\right) \parallel \leq \gamma \parallel x\parallel + c \) . (b) In the presence of Standing Hypotheses 1.2(a, c), property (b) is equivalent to the graph of \( F \) being closed. (c) Let \( \Omega \) be a bounded subset of \( \mathbb{R} \times {\mathbb{R}}^{n} \) . Prove that \( F \) is uniformly bounded on \( \Omega \) : there exists \( M \) such that \[ \left( {t, x}\right) \in \Omega ,\;v \in F\left( {t, x}\right) \Rightarrow \parallel v\parallel \leq M. \] (d) Let \( \Omega \) be a bounded subset of \( \mathbb{R} \times {\mathbb{R}}^{n} \), and let \( r > 0 \) satisfy \( \parallel \left( {t, x}\right) \parallel \leq r\forall \left( {t, x}\right) \in \Omega \) . Let \( \widetilde{r} > r \), and define \( \widetilde{F} \) as follows: \[ \widetilde{F}\left( {t, x}\right) = \left\{ \begin{array}{ll} F\left( {t, x}\right) & \text{ if }\parallel \left( {t, x}\right) \parallel \leq \widetilde{r}, \\ F\left( {\frac{\left( t, x\right) }{\parallel \left( {t, x}\right) \parallel }\widetilde{r}}\right) & \text{ if }\parallel \left( {t, x}\right) \parallel \geq \widetilde{r}. \end{array}\right. \] Prove that \( \widetilde{F} \) satisfies the Standing Hypotheses, agrees with \( F \) on a neighborhood of \( \Omega \), and is globally bounded. The role of the linear growth condition in the classical theory of differential equations is predicated on the a priori bounds on solutions to which it gives rise. We will benefit from it in precisely the same way. The following is known as Gronwall's Lemma. 1.4. Proposition. Let \( x \) be an arc on \( \left\lbrack {a, b}\right\rbrack \) which satisfies \[ \parallel \dot{x}\left( t\right) \parallel \leq \gamma \parallel x\left( t\right) \parallel + c\left( t\right) \text{ a.e.,}\;t \in \left\lbrack {a, b}\right\rbrack , \] where \( \gamma \) is a nonnegative constant and where \( c\left( \cdot \right) \in {L}_{1}^{1}\left\lbrack {a, b}\right\rbrack \) . Then, for all \( t \in \left\lbrack {a, b}\right\rbrack \), we have \[ \parallel x\left( t\right) - x\left( a\right) \parallel \leq \left( {{e}^{\gamma \left( {t - a}\right) } - 1}\right) \parallel x\left( a\right) \parallel + {\int }_{a}^{t}{e}^{\gamma \left( {t - s}\right) }c\left( s\right) {ds}. \] If the function \( c \) is constant and \( \gamma > 0 \), this becomes \[ \parallel x\left( t\right) - x\left( a\right) \parallel \leq \left( {{e}^{\gamma \left( {t - a}\right) } - 1}\right) \left( {\parallel x\left( a\right) \parallel + c/\gamma }\right) . \] Proof. Let \( r\left( t\right) \mathrel{\text{:=}} \parallel x\left( t\right) - x\left( a\right) \parallel \), a function which is absolutely continuous on \( \left\lbrack {a, b}\right\rbrack \), as the composition of a Lipschitz function and an absolutely continuous one. Let \( t \) be in that set of full measure in which both \( \dot{x}\left( t\right) \) and \( \dot{r}\left( t\right) \) exist. If \( x\left( t\right) \neq x\left( a\right) \), we have \[ \dot{r}\left( t\right) = \frac{\langle \dot{x}\left( t\right), x\left( t\right) - x\left( a\right) \rangle }{\parallel x\left( t\right) - x\left( a\right) \parallel } \] and otherwise \( \dot{r}\left( t\right) = 0 \) (since \( r \) attains a minimum at \( t \) ). Thus, a.e. \( t \in \left\lbrack {a, b}\right\rbrack \) we have \[ \dot{r}\left( t\right) \leq \parallel \dot{x}\left( t\right) \parallel \leq \gamma \parallel x\left( t\right) \parallel + c\left( t\right) \] \[ \leq \gamma \parallel x\left( t\right) - x\left( a\right) \parallel + \gamma \parallel x\left( a\right) \parallel + c\left( t\right) \] \[ = {\gamma r}\left( t\right) + \gamma \parallel x\left( a\right) \parallel + c\left( t\right) . \] We rewrite this inequality in the form \[ \left( {\dot{r}\left( t\right) - {\gamma r}\left( t\right) }\right) {e}^{-{\gamma t}} \leq \gamma {e}^{-{\gamma t}}\parallel x\left( a\right) \parallel + {e}^{-{\gamma t}}c\left( t\right) \] and note that the left side is the derivative of the function \( t \mapsto r\left( t\right) {e}^{-{\gamma t}} \) . Integrating both sides now gives the result. 1.5. Exercise. Let \( C \) be a bounded subset of \( {\mathbb{R}}^{n} \) and let \( \left\lbrack {a, b}\right\rbrack \) be given. Show that there exists \( K > 0 \) and \( M > 0 \) with the property that any trajectory \( x \) of \( F \) on \( \left\lbrack {a, b}\right\rbrack \) having \( x\left( a\right) \in C \) is Lipschitz on \( \left\lbrack {a, b}\right\rbrack \) of rank \( K \), and has \( \parallel x\left( t\right) \parallel \leq M \) for every \( t \in \left\lbrack {a, b}\right\rbrack \) . ## Euler Solutions Many of us will have seen methods of calculating solutions of ordinary differential equations; how would we study in concrete terms the calculation of trajectories of the differential inclusion (1)? The most straightforward approach to calculating a trajectory is to first find a selection \( f \) of \( F \) ; i.e., a function \( f \) such that \( f\left( {t, x}\right) \in F\left( {t, x}\right) \) for all \( \left( {t, x}\right) \) . Then, we consider the differential equation \( \dot{x} = f\left( {t, x}\right) \) ; any solution will presumably satisfy (1). The problem with this approach lies in finding selections \( f \) with the regularity properties (e.g., continuity) required by the usual theory of differential equations. This selection issue is an interesting and well-studied one, but not one that we intend nor need to dwell upon. Instead, we will consider a generalized concept of solution to \( \dot{x} = f\left( {t, x}\right) \), one which requires no particular regularity of \( f \) . Let us now consider the so-called Cauchy or initial-value problem \[ \dot{x}\left( t\right) = f\left( {t, x\left( t\right) }\right) ,\;x\left( a\right) = {x}_{0}, \] (4) where \( f \) is simply any function from \( \left\lbrack {a, b}\right\rbrack \times {\mathbb{R}}^{n} \) to \( {\mathbb{R}}^{n} \) . (We will return presently to our differential inclusion.) How would we begin to calculate numerically a solution of (4)? Recalling the classical Euler iterative scheme from ordinary differential equations, we suspect that a reasonable answer is obtained by discretizing in time. So let \[ \pi = \left\{ {{t}_{0},{t}_{1},\ldots ,{t}_{N - 1},{t}_{N}}\right\} \] be a partition of \( \left\lbrack {a, b}\right\rbrack \), where \( {t}_{0} = a \) and \( {t}_{N} = b \) . (We do not require uniform partitions; thus the interval lengths \( {t}_{i} - {t}_{i - 1} \) may differ.) We proceed by considering, on the interval \( \left\lbrack {{t}_{0},{t}_{1}}\right\rbrack \), the differential equation with constant right-hand side \[ \dot{x}\left( t\right) = f\left( {{t}_{0},{x}_{0}}\right) ,\;x\left( {t}_{0}\right) = {x}_{0}. \] Of course this has a unique solution \( x\left( t\right) \) on \( \left\lbrack {{t}_{0},{t}_{1}}\right\rbrack \), since the right side is constant; we define \( {x}_{1} \mathrel{\text{:=}} x\left( {t}_{1}\right) \) . Next we iterate, by considering on \( \left\lbrack {{t}_{1},{t}_{2}}\right\rbrack \) the initial-value problem \[ \dot{x}\left( t\right) = f\left( {{t}_{1},{x}_{1}}\right) ,\;x\left( {t}_{1}\right) = {x}_{1}. \] The next so-called node of the scheme is \( {x}_{2} \mathrel{\text{:=}} x\left( {t}_{2}\right) \) . We proceed in this manner until an arc \( {x}_{\pi } \) (which is in fact piecewise affine) has been defined on all of \( \left\lbrack {a, b}\right\rbrack \) . We use the notation \( {x}_{\pi } \) to emphasize the role played by the particular partition \( \pi \) in determining \( {x}_{\pi } \), which has been called in the past (and in our present) the Euler polygonal arc corresponding to the partition \( \pi \), or similar words to that effect. The diamet
1012_(GTM178)Nonsmooth Analysis and Control Theory
64
}. \] Of course this has a unique solution \( x\left( t\right) \) on \( \left\lbrack {{t}_{0},{t}_{1}}\right\rbrack \), since the right side is constant; we define \( {x}_{1} \mathrel{\text{:=}} x\left( {t}_{1}\right) \) . Next we iterate, by considering on \( \left\lbrack {{t}_{1},{t}_{2}}\right\rbrack \) the initial-value problem \[ \dot{x}\left( t\right) = f\left( {{t}_{1},{x}_{1}}\right) ,\;x\left( {t}_{1}\right) = {x}_{1}. \] The next so-called node of the scheme is \( {x}_{2} \mathrel{\text{:=}} x\left( {t}_{2}\right) \) . We proceed in this manner until an arc \( {x}_{\pi } \) (which is in fact piecewise affine) has been defined on all of \( \left\lbrack {a, b}\right\rbrack \) . We use the notation \( {x}_{\pi } \) to emphasize the role played by the particular partition \( \pi \) in determining \( {x}_{\pi } \), which has been called in the past (and in our present) the Euler polygonal arc corresponding to the partition \( \pi \), or similar words to that effect. The diameter (or mesh size) \( {\mu }_{\pi } \) of the partition \( \pi \) is given by \[ {\mu }_{\pi } \mathrel{\text{:=}} \max \left\{ {{t}_{i} - {t}_{i - 1} : 1 \leq i \leq N}\right\} . \] An Euler solution to the initial-value problem (4) means any arc \( x \) which is the uniform limit of Euler polygonal arcs \( {x}_{{\pi }_{j}} \), corresponding as above to some sequence \( {\pi }_{j} \) such that \( {\pi }_{j} \rightarrow 0 \), where this connotes convergence of the diameters \( {\mu }_{{\pi }_{j}} \rightarrow 0 \) (evidently, the corresponding number \( {N}_{j} \) of partition points in \( {\pi }_{j} \) must then go to infinity). We will also say that an arc \( x \) on \( \left\lbrack {a, b}\right\rbrack \) is an Euler arc for \( f \) when \( x \) is an Euler solution on \( \left\lbrack {a, b}\right\rbrack \) as above to the initial-value problem (4) for the "right" initial condition, namely \( {x}_{0} = x\left( a\right) \) . There are potential pathologies associated with these Euler solutions when \( f \) is discontinuous, one of which is that the equality \( \dot{x}\left( t\right) = f\left( {t, x\left( t\right) }\right) \) may fail completely. The following exercise serves to illustrate this, among other perhaps counterintuitive and unexpected features. 1.6. Exercise. (a) Define \( f : \left\lbrack {0,1}\right\rbrack \times \mathbb{R} \rightarrow \mathbb{R} \) as follows: \[ f\left( {t, x}\right) = \left\{ \begin{array}{ll} {e}^{t} & \text{ if }x = {e}^{t} \\ 1 & \text{ if }x < {e}^{t} \\ e & \text{ if }x > {e}^{t} \end{array}\right. \] Let \( {x}_{0} = 1, a = 0 \), and \( b = 1 \) . Show that problem (4) admits a unique Euler solution but that the arc \( \widehat{x}\left( t\right) = {e}^{t} \) is not an Euler solution, even though it satisfies \( \dot{x}\left( t\right) = f\left( {t, x\left( t\right) }\right) \) at every point. (b) Let \( f \) be the function \[ f\left( x\right) = \begin{cases} 1 & \text{ if }x \leq 0 \\ - 1 & \text{ if }x > 0 \end{cases} \] Set \( {x}_{0} = 0, a = 0 \), and \( b = 1 \) . Show that \( x\left( t\right) \equiv 0 \) is the unique Euler solution of (4), although it fails to satisfy \( \dot{x}\left( t\right) = f\left( {t, x\left( t\right) }\right) \) at any \( t \) . (c) Now let \( f \) be the function \[ f\left( x\right) = \begin{cases} 1 & \text{ if }x = 0 \\ - 1 & \text{ otherwise } \end{cases} \] where again \( {x}_{0} = 0, a = 0 \), and \( b = 1 \) . Show that \( x\left( t\right) \equiv 0 \) and \( x\left( t\right) = - t \) are both Euler solutions of the initial-value problem. (Hint. The first Euler solution is arrived at by considering only uniform partitions.) (d) Use part (c) in order to illustrate the fact that if we are restricted to exclusively uniform partitions of a compact interval \( \left\lbrack {a, b}\right\rbrack \), then the set of Euler solutions obtained thereby can be a proper subset of the general class of Euler solutions (even when \( f \) is dependent only upon \( x \) ). (e) Provide an example of a function \( f : \mathbb{R} \rightarrow \mathbb{R} \) such that there exist two distinct Euler solutions of (4) on \( \left\lbrack {0,1}\right\rbrack \), each corresponding to a sequence of uniform partitions. On the other hand, here are some important positive things that can be said about Euler solutions: 1.7. Theorem. Suppose that for positive constants \( \gamma \) and \( c \) and for all \( \left( {t, x}\right) \in \left\lbrack {a, b}\right\rbrack \times {\mathbb{R}}^{n} \), we have the linear growth condition \[ \parallel f\left( {t, x}\right) \parallel \leq \gamma \parallel x\parallel + c \] where \( f \) is otherwise arbitrary. Then: (a) At least one Euler solution \( x \) to the initial-value problem (4) exists on \( \left\lbrack {a, b}\right\rbrack \), and any Euler solution is Lipschitz. (b) Any Euler arc \( x \) for \( f \) on \( \left\lbrack {a, b}\right\rbrack \) satisfies \[ \parallel x\left( t\right) - x\left( a\right) \parallel \leq \left( {t - a}\right) {e}^{\gamma \left( {t - a}\right) }\left( {c + \gamma \parallel x\left( a\right) \parallel }\right) ,\;a \leq t \leq b. \] (c) If \( f \) is continuous, then any Euler arc \( x \) for \( f \) on \( \left\lbrack {a, b}\right\rbrack \) is continuously differentiable on \( \left( {a, b}\right) \) and satisfies \( \dot{x}\left( t\right) = f\left( {t, x\left( t\right) }\right) \forall t \in \left( {a, b}\right) \) . Proof. Let \( \pi \mathrel{\text{:=}} \left\{ {{t}_{0},{t}_{1},\ldots ,{t}_{N}}\right\} \) be a partition of \( \left\lbrack {a, b}\right\rbrack \), and let \( {x}_{\pi } \) be the corresponding Euler polygonal arc, with the nodes of \( {x}_{\pi } \) being denoted \( {x}_{0},{x}_{1},\ldots ,{x}_{N} \) as usual. On the interval \( \left( {{t}_{i},{t}_{i + 1}}\right) \) we have \[ \begin{Vmatrix}{{\dot{x}}_{\pi }\left( t\right) }\end{Vmatrix} = \begin{Vmatrix}{f\left( {{t}_{i},{x}_{i}}\right) }\end{Vmatrix} \leq \gamma \begin{Vmatrix}{x}_{i}\end{Vmatrix} + c \] whence \[ \begin{Vmatrix}{{x}_{i + 1} - {x}_{0}}\end{Vmatrix} \leq \begin{Vmatrix}{{x}_{i + 1} - {x}_{i}}\end{Vmatrix} + \begin{Vmatrix}{{x}_{i} - {x}_{0}}\end{Vmatrix} \] \[ \leq \left( {{t}_{i + 1} - {t}_{i}}\right) \left( {\gamma \begin{Vmatrix}{x}_{i}\end{Vmatrix} + c}\right) + \begin{Vmatrix}{{x}_{i} - {x}_{0}}\end{Vmatrix} \] \[ \leq \left\lbrack {\left( {{t}_{i + 1} - {t}_{i}}\right) \gamma + 1}\right\rbrack \begin{Vmatrix}{{x}_{i} - {x}_{0}}\end{Vmatrix} + \left( {{t}_{i + 1} - {t}_{i}}\right) \left( {\gamma \begin{Vmatrix}{x}_{0}\end{Vmatrix} + c}\right) . \] We now require the following exercise in induction: 1.8. Exercise. Let \( {r}_{0},{r}_{1},\ldots ,{r}_{N} \) be nonnegative numbers satisfying \[ {r}_{i + 1} \leq \left( {1 + {\delta }_{i}}\right) {r}_{i} + {\Delta }_{i},\;i = 0,1,\ldots, N - 1, \] where \( {\delta }_{i} \geq 0 \) and \( {\Delta }_{i} \geq 0,{r}_{0} = 0 \) . Then \[ {r}_{N} \leq \left( {\exp \left( {\mathop{\sum }\limits_{{i = 0}}^{{N - 1}}{\delta }_{i}}\right) }\right) \mathop{\sum }\limits_{{i = 0}}^{{N - 1}}{\Delta }_{i} \] Returning now to the authors' part of the task, we apply this result to derive, for \( i = 1,2,\ldots, N \) , \[ \begin{Vmatrix}{{x}_{i} - {x}_{0}}\end{Vmatrix} \leq M \] where \[ M \mathrel{\text{:=}} \left( {b - a}\right) {e}^{\gamma \left( {b - a}\right) }\left( {\gamma \begin{Vmatrix}{x}_{0}\end{Vmatrix} + c}\right) . \] Thus all the nodes \( {x}_{i} \) lie in the ball \( \bar{B}\left( {{x}_{0};M}\right) \) ; by convexity this is true of all the values \( {x}_{\pi }\left( t\right), a \leq t \leq b \) . Since the derivative along any linear portion of \( {x}_{\pi } \) is determined by the values of \( f \) at the nodes, we obtain as well the following uniform bound on \( \left\lbrack {a, b}\right\rbrack \) : \[ {\begin{Vmatrix}{\dot{x}}_{\pi }\end{Vmatrix}}_{\infty } \leq k \mathrel{\text{:=}} {\gamma M} + c. \] Therefore \( {x}_{\pi } \) is Lipschitz of rank \( k \) on \( \left\lbrack {a, b}\right\rbrack \) . Now let \( {\pi }_{j} \) be a sequence of partitions such that \( {\pi }_{j} \rightarrow 0 \) ; i.e., such that \( {\mu }_{{\pi }_{j}} \) goes to zero, and (necessarily) \( {N}_{j} \rightarrow \infty \) . Then the corresponding polygonal \( \operatorname{arcs}{x}_{{\pi }_{j}} \) on \( \left\lbrack {a, b}\right\rbrack \) all satisfy \[ {x}_{{\pi }_{j}}\left( a\right) = {x}_{0},\;{\begin{Vmatrix}{x}_{{\pi }_{j}} - {x}_{0}\end{Vmatrix}}_{\infty } \leq M,\;{\begin{Vmatrix}{\dot{x}}_{{\pi }_{j}}\end{Vmatrix}}_{\infty } \leq k. \] It follows that the family \( \left\{ {x}_{{\pi }_{j}}\right\} \) is equicontinuous and uniformly bounded; then, by the well-known theorem of Arzela and Ascoli, some subsequence of it converges uniformly to a continuous function \( x \) . The limiting function inherits the Lipschitz rank \( k \) on \( \left\lbrack {a, b}\right\rbrack \), and in consequence is absolutely continuous (i.e., \( x \) is an arc). Thus by definition \( x \) is an Euler solution of the initial-value problem (4) on \( \left\lbrack {a, b}\right\rbrack \), and assertion (a) of the theorem is proved. The inequality in part (b) of the theorem is inherited by \( x \) from the sequence of polygonal arcs generating it (we identify \( t \) with \( b \) ). There remains to prove part (c) of the theorem. To this end, let \( {x}_{{\pi }_{j}} \) denote a sequence of polygonal arcs for problem (4) converging uniformly to an Euler solution \( x \) . As shown above, the arcs \( {x}_{{\pi }_{i}} \) all lie in a certain ball \( \bar{B}\left( {{x}_{0};M}\right) \) and they all satisfy a Lipschitz condition of the same rank \( k \) . Since a continuous function on \( {\mathbb{R}}^{n} \) is uniformly continuous on compact sets, for any \( \varepsilon > 0 \), we can find \( \delta > 0 \) such that \[ t,\widetilde{t} \in \left\lbrack {a, b}\right\rbrack ,\;x,\widetilde{x} \in \bar{B}\left( {{x}_{0};M}\right) ,\;\left| {t - \widetilde{t}}\right| < \delta , \] \[ \parall
1012_(GTM178)Nonsmooth Analysis and Control Theory
65
solution of the initial-value problem (4) on \( \left\lbrack {a, b}\right\rbrack \), and assertion (a) of the theorem is proved. The inequality in part (b) of the theorem is inherited by \( x \) from the sequence of polygonal arcs generating it (we identify \( t \) with \( b \) ). There remains to prove part (c) of the theorem. To this end, let \( {x}_{{\pi }_{j}} \) denote a sequence of polygonal arcs for problem (4) converging uniformly to an Euler solution \( x \) . As shown above, the arcs \( {x}_{{\pi }_{i}} \) all lie in a certain ball \( \bar{B}\left( {{x}_{0};M}\right) \) and they all satisfy a Lipschitz condition of the same rank \( k \) . Since a continuous function on \( {\mathbb{R}}^{n} \) is uniformly continuous on compact sets, for any \( \varepsilon > 0 \), we can find \( \delta > 0 \) such that \[ t,\widetilde{t} \in \left\lbrack {a, b}\right\rbrack ,\;x,\widetilde{x} \in \bar{B}\left( {{x}_{0};M}\right) ,\;\left| {t - \widetilde{t}}\right| < \delta , \] \[ \parallel x - \widetilde{x}\parallel < \delta \Rightarrow \parallel f\left( {t, x}\right) - f\left( {\widetilde{t},\widetilde{x}}\right) \parallel < \varepsilon . \] Now let \( j \) be large enough so that the partition diameter \( {\mu }_{{\pi }_{j}} \) satisfies \( {\mu }_{{\pi }_{j}} < \delta \) and \( k{\mu }_{{\pi }_{j}} < \delta \) . For any point \( t \) which is not one of the finitely many points at which \( {x}_{{\pi }_{j}}\left( t\right) \) is a node, we have \( {\dot{x}}_{{\pi }_{j}}\left( t\right) = f\left( {\widetilde{t},{x}_{{\pi }_{j}}\left( \widetilde{t}\right) }\right) \) for some \( \widetilde{t} \) within \( {\mu }_{{\pi }_{j}} < \delta \) of \( t \) . Thus, since \[ \begin{Vmatrix}{{x}_{{\pi }_{j}}\left( t\right) - {x}_{{\pi }_{j}}\left( \widetilde{t}\right) }\end{Vmatrix} \leq k{\mu }_{{\pi }_{j}} < \delta \] we deduce \[ \begin{Vmatrix}{{\dot{x}}_{{\pi }_{j}}\left( t\right) - f\left( {t,{x}_{{\pi }_{j}}\left( t\right) }\right) }\end{Vmatrix} = \begin{Vmatrix}{f\left( {t,{x}_{{\pi }_{j}}\left( t\right) }\right) - f\left( {\widetilde{t},{x}_{{\pi }_{j}}\left( \widetilde{t}\right) }\right) }\end{Vmatrix} < \varepsilon . \] It follows that for any \( t \) in \( \left\lbrack {a, b}\right\rbrack \), we have \[ \begin{Vmatrix}{{x}_{{\pi }_{j}}\left( t\right) - {x}_{{\pi }_{j}}\left( a\right) - {\int }_{a}^{t}f\left( {t,{x}_{{\pi }_{j}}\left( \tau \right) }\right) {d\tau }}\end{Vmatrix} \] \[ = \begin{Vmatrix}{{\int }_{a}^{t}\left\{ {{\dot{x}}_{{\pi }_{j}}\left( \tau \right) - f\left( {\tau ,{x}_{{\pi }_{j}}\left( \tau \right) }\right) }\right\} {d\tau }}\end{Vmatrix} < \varepsilon \left( {t - a}\right) \leq \varepsilon \left( {b - a}\right) . \] Letting \( j \rightarrow \infty \), we obtain from this \[ \begin{Vmatrix}{x\left( t\right) - {x}_{0} - {\int }_{a}^{t}f\left( {\tau, x\left( \tau \right) }\right) {d\tau }}\end{Vmatrix} \leq \varepsilon \left( {b - a}\right) . \] Since \( \varepsilon \) is arbitrary, it follows that \[ x\left( t\right) = {x}_{0} + {\int }_{a}^{t}f\left( {\tau, x\left( \tau \right) }\right) {d\tau } \] which implies (since the integrand is continuous) that \( x \) is \( {C}^{1} \) and \( \dot{x}\left( t\right) = \) \( f\left( {t, x\left( t\right) }\right) \) for all \( t \in \left( {a, b}\right) \) . Euler arcs possess a uniform Lipschitz property akin to the one established in Exercise 1.5 for trajectories: 1.9. Exercise. Let \( C \) be a bounded subset of \( {\mathbb{R}}^{n} \) and let \( \left\lbrack {a, b}\right\rbrack \) be given. Assume that \( f \) is given as in Theorem 1.7. Show that there exists \( K > 0 \) with the property that any Euler arc \( x \) for \( f \) on \( \left\lbrack {a, b}\right\rbrack \) having \( x\left( a\right) \in C \) is Lipschitz on \( \left\lbrack {a, b}\right\rbrack \) of rank \( K \) . We are reassured by part (c) of the theorem that when \( f \) is continuous (which is the minimal assumption under which the classical study of differential equations operates), Euler arcs satisfy the usual pointwise definition of solution. Part (a) of the following exercise will provide a converse, and in part (b), a further counterintuitive feature is noted: 1.10. Exercise. (a) Let \( f \) be locally Lipschitz, and let the arc \( x \) on \( \left\lbrack {a, b}\right\rbrack \) satisfy \( \dot{x}\left( t\right) = f\left( {t, x\left( t\right) }\right) \forall t \in \left( {a, b}\right) \) . Prove that \( x \) is an Euler arc for \( f \) on \( \left\lbrack {a, b}\right\rbrack \) . (b) Let \( f\left( {t, x}\right) \mathrel{\text{:=}} \left( {3/2}\right) {x}^{1/3} \) (for \( n = 1 \) ), \( {x}_{0} = 0, a = 0, b = 1 \) . Prove that the initial-value problem (4) has three distinct classical solutions of the form \( x\left( t\right) = \alpha {t}^{\beta } \), but only one Euler solution. ## Compactness of Approximate Trajectories We now return to considering the trajectories of a multifunction \( F \) satisfying the Standing Hypotheses, and specifically the existence issue for differential inclusions. Let \( f \) be any selection for \( F \) ; that is, \[ f\left( {t, x}\right) \in F\left( {t, x}\right) \forall \left( {t, x}\right) . \] Then \( f \) evidently inherits the linear growth condition from \( F \) . By Theorem 1.7, an Euler solution \( x \) to the initial-value problem (4) exists. We are tempted to conclude that \( \dot{x}\left( t\right) \), being equal to \( f\left( {t, x\left( t\right) }\right) \) a.e., must lie in \( F\left( {t, x\left( t\right) }\right) \) a.e.; i.e., that \( x \) is a trajectory for \( F \) . This reasoning is fallacious, however, since an Euler solution to (4) may not satisfy the pointwise condition \( \dot{x} = f\left( {t, x}\right) \) for any \( t \) ; witness Exercise 1.6(b). The key to the correct argument is a sequential compactness property of approximate trajectories, one which we will have frequent occasion to invoke, and which we establish in some generality for future purposes. 1.11. Theorem. Let \( \left\{ {x}_{i}\right\} \) be a sequence of arcs on \( \left\lbrack {a, b}\right\rbrack \) such that the set \( \left\{ {{x}_{i}\left( a\right) }\right\} \) is bounded, and satisfying \[ {\dot{x}}_{i}\left( t\right) \in F\left( {{\tau }_{i}\left( t\right) ,{x}_{i}\left( t\right) + {y}_{i}\left( t\right) }\right) + {r}_{i}\left( t\right) \text{B a.e.,} \] where \( \left\{ {y}_{i}\right\} ,\left\{ {r}_{i}\right\} ,\left\{ {\tau }_{i}\right\} \) are sequences of measurable functions on \( \left\lbrack {a, b}\right\rbrack \) such that \( {y}_{i} \) converges to 0 in \( {L}^{2},{r}_{i} \geq 0 \) converges to 0 in \( {L}^{2},{\tau }_{i} \) converges a.e. to \( t \) . Then there is a subsequence of \( \left\{ {x}_{i}\right\} \) which converges uniformly to an arc \( x \) which is a trajectory of \( F \), and whose derivatives converge weakly to \( \dot{x} \) . Proof. From the differential inclusion and the linear growth condition we have \[ \begin{Vmatrix}{{\dot{x}}_{i}\left( t\right) }\end{Vmatrix} \leq \gamma \begin{Vmatrix}{{x}_{i}\left( t\right) + {y}_{i}\left( t\right) }\end{Vmatrix} + \left| {{r}_{i}\left( t\right) }\right| . \] Using Gronwall's Lemma (Proposition 1.4), this implies a uniform bound on \( {\begin{Vmatrix}{x}_{i}\end{Vmatrix}}_{\infty } \) and hence on \( {\begin{Vmatrix}{\dot{x}}_{i}\end{Vmatrix}}_{2} \) . Invoking weak compactness in \( {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) allows the extraction of a subsequence \( {\dot{x}}_{{i}_{j}} \) converging weakly to a limit \( {v}_{0} \) ; we may also suppose (by Arzela and Ascoli) that \( {x}_{{i}_{j}} \) converges uniformly to a continuous function \( x \) . Passing to the limit in \[ {x}_{{i}_{j}}\left( t\right) = {x}_{{i}_{j}}\left( a\right) + {\int }_{a}^{t}{\dot{x}}_{{i}_{j}}\left( s\right) {ds} \] shows that \( x\left( t\right) = x\left( a\right) + {\int }_{a}^{t}{v}_{0}\left( s\right) {ds} \), whence \( x \) is an arc and \( \dot{x} = {v}_{0} \) a.e. The fact that \( x \) is a trajectory for \( F \) is an immediate consequence of Theorem 3.5.24. The following important consequence of the theorem vindicates our attempt to calculate trajectories of \( F \) by way of selections, and proves that trajectories do exist, given any \( {x}_{0} \) and \( \left\lbrack {a, b}\right\rbrack \) . 1.12. Corollary. Let \( f \) be any selection of \( F \), and let \( x \) be an Euler solution on \( \left\lbrack {a, b}\right\rbrack \) of \( \dot{x} = f\left( {t, x}\right), x\left( a\right) = {x}_{0} \) . Then \( x \) is a trajectory of \( F \) on \( \left\lbrack {a, b}\right\rbrack \) . Proof. Let \( {x}_{{\pi }_{j}} \) be the polygonal arcs whose uniform limit is \( x \), as in the proof of Theorem 1.7. Let \( t \in \left( {a, b}\right) \) be a nonpartition point, and let \( {\tau }_{j}\left( t\right) \) designate the partition point \( {t}_{i} \) immediately before \( t \) . Then \[ {\dot{x}}_{{\pi }_{j}}\left( t\right) = f\left( {{t}_{i},{x}_{i}}\right) \in F\left( {{t}_{i},{x}_{i}}\right) = F\left( {{\tau }_{j}\left( t\right) ,{x}_{{\pi }_{j}}\left( t\right) + {y}_{j}\left( t\right) }\right) , \] where \( {y}_{j}\left( t\right) \mathrel{\text{:=}} {x}_{i} - {x}_{{\pi }_{j}}\left( t\right) = {x}_{{\pi }_{j}}\left( {{\tau }_{j}\left( t\right) }\right) - {x}_{{\pi }_{j}}\left( t\right) \) . Since the functions \( {x}_{{\pi }_{j}} \) admit a common Lipschitz constant \( k \), we have \[ {\begin{Vmatrix}{y}_{j}\left( t\right) \end{Vmatrix}}_{\infty } \leq k\mathop{\sup }\limits_{{t \in \left\lbrack {a, b}\right\rbrack }}\left| {{\tau }_{j}\left( t\right) - t}\right| \leq k{\mu }_{{\pi }_{j}}. \] It follows that \( {\tau }_{j} \) and \( {y}_{j} \) are (measurable) functions converging uniformly to \( t \) and 0, respectively. It is now a consequence of the theorem that \( x \), the uniform limit of \( {x}_{{\pi }_{j}} \), is a trajectory of \( F \) . ## 1.13. Exercise. (a) Trajectory continuation. Let \( z \) be a trajectory of \( F \) on \( \left\lbrac
1012_(GTM178)Nonsmooth Analysis and Control Theory
66
,{x}_{i}}\right) = F\left( {{\tau }_{j}\left( t\right) ,{x}_{{\pi }_{j}}\left( t\right) + {y}_{j}\left( t\right) }\right) , \] where \( {y}_{j}\left( t\right) \mathrel{\text{:=}} {x}_{i} - {x}_{{\pi }_{j}}\left( t\right) = {x}_{{\pi }_{j}}\left( {{\tau }_{j}\left( t\right) }\right) - {x}_{{\pi }_{j}}\left( t\right) \) . Since the functions \( {x}_{{\pi }_{j}} \) admit a common Lipschitz constant \( k \), we have \[ {\begin{Vmatrix}{y}_{j}\left( t\right) \end{Vmatrix}}_{\infty } \leq k\mathop{\sup }\limits_{{t \in \left\lbrack {a, b}\right\rbrack }}\left| {{\tau }_{j}\left( t\right) - t}\right| \leq k{\mu }_{{\pi }_{j}}. \] It follows that \( {\tau }_{j} \) and \( {y}_{j} \) are (measurable) functions converging uniformly to \( t \) and 0, respectively. It is now a consequence of the theorem that \( x \), the uniform limit of \( {x}_{{\pi }_{j}} \), is a trajectory of \( F \) . ## 1.13. Exercise. (a) Trajectory continuation. Let \( z \) be a trajectory of \( F \) on \( \left\lbrack {a, b}\right\rbrack \) . Prove that there exists a trajectory of \( F \) on \( \lbrack a,\infty ) \) which coincides with \( z \) on \( \left\lbrack {a, b}\right\rbrack \) . What is more, show that there exists a trajectory of \( F \) on \( \left( {-\infty ,\infty }\right) \) which coincides with \( z \) on \( \left\lbrack {a, b}\right\rbrack \) . (b) Variable intervals. Let \( {x}_{i} \) be a sequence of trajectories of \( F \) on \( \left\lbrack {{a}_{i},{b}_{i}}\right\rbrack \), where \( {a}_{i} \rightarrow a,{b}_{i} \rightarrow b, a < b \), and where the sequence \( {x}_{i}\left( {a}_{i}\right) \) is bounded. Let the trajectories \( {x}_{i} \) be continued to \( \left( {-\infty ,\infty }\right) \) as in part (a). Prove that a subsequence \( \left\{ {x}_{{i}_{j}}\right\} \) of \( \left\{ {x}_{i}\right\} \) has the property that, for some trajectory \( \bar{x} \) of \( F \) on \( \left\lbrack {a, b}\right\rbrack \) , \( {x}_{{i}_{j}} \) converges uniformly to \( \bar{x} \) on \( \left\lbrack {a, b}\right\rbrack \) . (c) Uniform convergence on bounded intervals. Let \( {x}_{i} \) be a sequence of trajectories of \( F \) on \( \lbrack a,\infty ) \) such that the sequence \( {x}_{i}\left( a\right) \) is bounded. Prove that there is a trajectory \( \bar{x} \) of \( F \) on \( \lbrack a,\infty ) \) and a subsequence \( {x}_{{i}_{j}} \) having the property that for any \( b > a,{x}_{{i}_{j}} \) converges uniformly to \( \bar{x} \) on \( \left\lbrack {a, b}\right\rbrack \) . (d) Let \( f : {\mathbb{R}}^{k} \rightarrow {\mathbb{R}}^{\ell } \) only be required to satisfy a linear growth condition. Is there a minimal multifunction \( F \) from \( {\mathbb{R}}^{k} \) to \( {\mathbb{R}}^{\ell } \) which satisfies the Standing Hypotheses such that \( f\left( x\right) \in F\left( x\right) \) for all \( x \) ? (e) Let \( f : {\mathbb{R}}^{n} \times {\mathbb{R}}^{m} \rightarrow {\mathbb{R}}^{n} \) and \( g : {\mathbb{R}}^{n} \times {\mathbb{R}}^{m} \rightarrow {\mathbb{R}}^{m} \) satisfy a linear growth condition, and let \( \left( {x, y}\right) \) be an Euler arc for \( \left( {f, g}\right) \) on \( \left\lbrack {a, b}\right\rbrack \) . Suppose that \( f \) is continuous. Prove that \( x \) is \( {C}^{1} \) on \( \left( {a, b}\right) \) and satisfies \( \dot{x}\left( t\right) = f\left( {x\left( t\right), y\left( t\right) }\right) \) on \( \left( {a, b}\right) \) . We can extend the notion of Euler arcs for \( f \) from finite intervals to semi-infinite ones of the form \( \lbrack a,\infty ) \) as follows: we say that the arc \( x\left( \cdot \right) \) defined on \( \lbrack a,\infty ) \) is an Euler arc on \( \lbrack a,\infty ) \) provided that for any \( b \in \left( {a,\infty }\right) \), the arc \( x \) restricted to \( \left\lbrack {a, b}\right\rbrack \) is an Euler arc for \( f \) on \( \left\lbrack {a, b}\right\rbrack \) (or, equivalently, an Euler solution on \( \left\lbrack {a, b}\right\rbrack \) of the initial-value problem \( \dot{y} = f\left( {t, y}\right), y\left( a\right) = x\left( a\right) \) ). In general, it is not the case that the concatenation of two Euler arcs for \( f \) is an Euler arc. That is, if \( x \) on \( \left\lbrack {a, b}\right\rbrack \) and \( y \) on \( \left\lbrack {b, c}\right\rbrack \) are Euler arcs, the arc consisting of \( x \) followed by \( y \) may not be one (on \( \left\lbrack {a, c}\right\rbrack \) ). Nonetheless, we have ## 1.14. Exercise. (a) Let \( f : \lbrack a,\infty ) \times {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{n} \) have linear growth. For any \( {x}_{0} \), show that there is an Euler arc \( x \) for \( f \) on \( \lbrack a,\infty ) \) such that \( x\left( a\right) = \) \( {x}_{0} \) . (Hint. Construct an appropriate family of polygonal arcs defined on \( \lbrack a,\infty ) \), then a subsequence converging on \( \left\lbrack {a, a + 1}\right\rbrack \) , a further subsequence converging on \( \left\lbrack {a, a + 2}\right\rbrack \), and so on.) (b) Prove Corollary 1.12 when \( \lbrack a,\infty ) \) replaces \( \left\lbrack {a, b}\right\rbrack \) . The lower Hamiltonian \( h \) and upper Hamiltonian \( H \) corresponding to \( F \) will play an important role in what follows. These are functions from \( \mathbb{R} \times {\mathbb{R}}^{n} \times {\mathbb{R}}^{n} \) to \( \mathbb{R} \) defined as follows: \[ h\left( {t, x, p}\right) \mathrel{\text{:=}} \mathop{\min }\limits_{{v \in F\left( {t, x}\right) }}\langle p, v\rangle ,\;H\left( {t, x, p}\right) \mathrel{\text{:=}} \mathop{\max }\limits_{{v \in F\left( {t, x}\right) }}\langle p, v\rangle . \] We gather some basic properties of \( h \) and \( H \) in the following, under the Standing Hypotheses on \( F \) . (Proposition 2.1.3 on support functions is relevant here.) ## 1.15. Exercise. (a) \( h \) is lower semicontinuous in \( \left( {x, p}\right) \), and concave and continuous in \( p \) . (b) \( h \) is superadditive in \( p \) : \[ h\left( {t, x, p + q}\right) \geq h\left( {t, x, p}\right) + h\left( {t, x, q}\right) , \] and \( h\left( {t, x,0}\right) = 0 \) . (c) A given vector \( v \) belongs to \( F\left( {t, x}\right) \) iff \[ h\left( {t, x, p}\right) \leq \langle p, v\rangle \forall p \in {\mathbb{R}}^{n} \] \( v \) belongs to \( F\left( {t, x}\right) + r\bar{B} \) (where \( r \geq 0 \) ) iff \[ h\left( {t, x, p}\right) \leq r\parallel p\parallel + \langle p, v\rangle \forall p \in {\mathbb{R}}^{n}. \] (d) What are the counterparts of (a)-(c) in terms of \( H \) ? ## 2 Weak Invariance A venerable notion from the classical theory of dynamical systems is that of invariance. When the basic model consists of an autonomous ordinary differential equation \( \dot{x}\left( t\right) = f\left( {x\left( t\right) }\right) \) with locally Lipschitz right-hand side and a set \( S \subseteq {\mathbb{R}}^{n} \), then flow invariance of the pair \( \left( {S, f}\right) \) is the property that for every initial point \( {x}_{0} \) in \( S \), the (unique) trajectory emanating from \( x\left( 0\right) = {x}_{0} \) is defined on \( \lbrack 0,\infty ) \) and satisfies \( x\left( t\right) \in S \) for all \( t \geq 0 \) . In this section, we will study a generalization of this concept to situations wherein the differential equation is replaced by a differential inclusion (possibly nonautonomous). The largely geometric analysis undertaken here will have important ramifications later in the chapter. Once again, we will find it advantageous to consider first Euler arcs. ## Proximal Aiming Suppose that we are studying the flow of an ordinary differential equation \[ \dot{x}\left( t\right) = f\left( {t, x\left( t\right) }\right) ,\;x\left( 0\right) = {x}_{0}, \] with the issue being to determine whether the resulting trajectory \( x\left( t\right) \) approaches a given closed set \( S \) in \( {\mathbb{R}}^{n} \) . One natural way to test whether this is the case is to pick a point \( s \in {\operatorname{proj}}_{S}\left( {x\left( t\right) }\right) \) (for a given \( t \) ), and check the sign of the quantity \[ \langle f\left( {t, x\left( t\right) }\right), x - s\rangle \] If it is negative, \( \dot{x}\left( t\right) \) will "point toward \( s \) ," and hence certainly toward \( S \) . If this prevails at every point \( \left( {t, x\left( t\right) }\right) \), then the state \( x\left( t\right) \) should indeed "move toward \( S \) ." The next result, a simple and important one, confirms this heuristic observation in the general framework of Euler arcs. 2.1. Proposition. Let \( f \) satisfy the linear growth condition \[ \parallel f\left( {t, x}\right) \parallel \leq \gamma \parallel x\parallel + c\forall \left( {t, x}\right) , \] and let \( x\left( \cdot \right) \) be an Euler arc for \( f \) on \( \left\lbrack {a, b}\right\rbrack \) . Let \( \Omega \) be an open set containing \( x\left( t\right) \forall t \in \left\lbrack {a, b}\right\rbrack \), and suppose that every \( \left( {t, z}\right) \in \left\lbrack {a, b}\right\rbrack \times \Omega \) satisfies the following "proximal aiming" condition: there exists \( s \in {\operatorname{proj}}_{S}\left( z\right) \) such that \( \langle f\left( {t, z}\right), z - s\rangle \leq 0 \) . Then we have \[ {d}_{S}\left( {x\left( t\right) }\right) \leq {d}_{S}\left( {x\left( a\right) }\right) \forall t \in \left\lbrack {a, b}\right\rbrack . \] Proof. Let \( {x}_{\pi } \) be one polygonal arc in the sequence converging uniformly to \( x \), as per the definition of the Euler solution. As usual, we denote its node at \( {t}_{i} \) by \( {x}_{i}\left( {i = 0,1,\ldots, N}\right) \), and so \( {x}_{0} = x\left( a\right) \) . We may suppose that \( {x}_{\pi }\left( t\right) \) lies in \( \Omega \) for all \( t \in \left\lbrack {a, b}\right\rbrack \) . Accordingly, there exists for each \( i \) a point \( {s}_{i} \in {\operatorname{proj}}_{S}\left( {x}_{i}\right) \) such that \( \left\langle {f\left( {{t}_{i},{x}_{i}}\right) ,{x}_{i} - {s}_{i}}\right\rangle \leq 0 \) . Letting
1012_(GTM178)Nonsmooth Analysis and Control Theory
67
times \Omega \) satisfies the following "proximal aiming" condition: there exists \( s \in {\operatorname{proj}}_{S}\left( z\right) \) such that \( \langle f\left( {t, z}\right), z - s\rangle \leq 0 \) . Then we have \[ {d}_{S}\left( {x\left( t\right) }\right) \leq {d}_{S}\left( {x\left( a\right) }\right) \forall t \in \left\lbrack {a, b}\right\rbrack . \] Proof. Let \( {x}_{\pi } \) be one polygonal arc in the sequence converging uniformly to \( x \), as per the definition of the Euler solution. As usual, we denote its node at \( {t}_{i} \) by \( {x}_{i}\left( {i = 0,1,\ldots, N}\right) \), and so \( {x}_{0} = x\left( a\right) \) . We may suppose that \( {x}_{\pi }\left( t\right) \) lies in \( \Omega \) for all \( t \in \left\lbrack {a, b}\right\rbrack \) . Accordingly, there exists for each \( i \) a point \( {s}_{i} \in {\operatorname{proj}}_{S}\left( {x}_{i}\right) \) such that \( \left\langle {f\left( {{t}_{i},{x}_{i}}\right) ,{x}_{i} - {s}_{i}}\right\rangle \leq 0 \) . Letting \( k \) be the a priori bound on \( {\begin{Vmatrix}{\dot{x}}_{\pi }\end{Vmatrix}}_{\infty } \) as in the proof of Theorem 1.7, we calculate \[ {d}_{S}^{2}\left( {x}_{1}\right) \leq {\begin{Vmatrix}{x}_{1} - {s}_{0}\end{Vmatrix}}^{2}\;\left( {\text{ since }{s}_{0} \in S}\right) \] \[ = {\begin{Vmatrix}{x}_{1} - {x}_{0}\end{Vmatrix}}^{2} + {\begin{Vmatrix}{x}_{0} - {s}_{0}\end{Vmatrix}}^{2} + 2\left\langle {{x}_{1} - {x}_{0},{x}_{0} - {s}_{0}}\right\rangle \] \[ \leq {k}^{2}{\left( {t}_{1} - {t}_{0}\right) }^{2} + {d}_{S}^{2}\left( {x}_{0}\right) + 2{\int }_{{t}_{0}}^{{t}_{1}}\left\langle {{\dot{x}}_{\pi }\left( t\right) ,{x}_{0} - {s}_{0}}\right\rangle {dt} \] \[ = {k}^{2}{\left( {t}_{1} - {t}_{0}\right) }^{2} + {d}_{S}^{2}\left( {x}_{0}\right) + 2{\int }_{{t}_{0}}^{{t}_{1}}\left\langle {f\left( {{t}_{0},{x}_{0}}\right) ,{x}_{0} - {s}_{0}}\right\rangle {dt} \] \[ \leq {k}^{2}{\left( {t}_{1} - {t}_{0}\right) }^{2} + {d}_{S}^{2}\left( {x}_{0}\right) \] The same estimates at any node apply to give \[ {d}_{S}^{2}\left( {x}_{i}\right) \leq {d}_{S}^{2}\left( {x}_{i - 1}\right) + {k}^{2}{\left( {t}_{i} - {t}_{i - 1}\right) }^{2} \] whence \[ {d}_{S}^{2}\left( {x}_{i}\right) \leq {d}_{S}^{2}\left( {x}_{0}\right) + {k}^{2}\mathop{\sum }\limits_{{\ell = 1}}^{i}{\left( {t}_{\ell } - {t}_{\ell - 1}\right) }^{2} \] \[ \leq {d}_{S}^{2}\left( {x}_{0}\right) + {k}^{2}{\mu }_{\pi }\mathop{\sum }\limits_{{\ell = 1}}^{i}\left( {{t}_{\ell } - {t}_{\ell - 1}}\right) \] \[ \leq {d}_{S}^{2}\left( {x}_{0}\right) + {k}^{2}{\mu }_{\pi }\left( {b - a}\right) \] Now consider the sequence \( {x}_{{\pi }_{j}} \) of polygonal arcs converging to \( x \) . Since the last estimate holds at every node and since \( {\mu }_{{\pi }_{j}} \rightarrow 0 \) (and the same \( k \) applies to each \( {x}_{{\pi }_{j}} \) ), we deduce in the limit \( {d}_{S}\left( {x\left( t\right) }\right) \leq {d}_{S}\left( {x\left( a\right) }\right) \forall t \in \left\lbrack {a, b}\right\rbrack \) , as claimed. 2.2. Exercise. Suppose that in Proposition 2.1, the proximal aim- ing condition is changed to \[ \langle f\left( {t, z}\right), z - s\rangle \leq \theta \left( {t, z}\right) {d}_{S}\left( z\right) \] for a continuous function \( \theta \), other things being equal. (a) Show that the conclusion becomes \[ {d}_{S}^{2}\left( {x\left( t\right) }\right) - {d}_{S}^{2}\left( {x\left( \tau \right) }\right) \leq 2{\int }_{\tau }^{t}\theta \left( {r, x\left( r\right) }\right) {d}_{S}\left( {x\left( r\right) }\right) {dr} \] for any \( a \leq \tau < t \leq b \) . (b) Prove that this implies \[ \frac{d}{dt}{d}_{S}\left( {x\left( t\right) }\right) \leq \theta \left( {t, x\left( t\right) }\right) \text{ a.e. } \] on any interval on which \( {d}_{S}\left( {x\left( t\right) }\right) > 0 \), or on any interval in which \( \theta \left( {t, x\left( t\right) }\right) \geq 0 \) . ## Weak Invariance (Autonomous Case) Our real goal is not to work with a given \( f\left( {t, x}\right) \), which allows no scope for control considerations, but rather with the possible trajectories of a differential inclusion \( \dot{x}\left( t\right) \in F\left( {x\left( t\right) }\right) \) . Note that we are taking \( F \) to be independent of \( t \) (the so-called autonomous case) for the moment; we will return to the nonautonomous case later. Our attention is still focused on approaching a given closed set \( S \) . More precisely, if the initial state \( {x}_{0} \) actually lies in \( S \) already, is there a trajectory starting at \( {x}_{0} \) that remains in \( S \) thereafter? This concept is important enough to merit some terminology: 2.3. Definition. The system \( \left( {S, F}\right) \) is called weakly invariant provided that for all \( {x}_{0} \in S \), there exists a trajectory \( x \) on \( \lbrack 0,\infty ) \) such that \[ x\left( 0\right) = {x}_{0},\;x\left( t\right) \in S\forall t \geq 0. \] Note that we speak of this property (which is also called viability) as being one of the pair \( \left( {S, F}\right) \), and not just of \( S \) . Here is a key proximal sufficient condition for weak invariance, in terms of the lower Hamiltonian defined earlier (see Exercise 1.15): 2.4. Theorem. Suppose that for every \( x \in S \), we have \[ h\left( {x,{N}_{S}^{P}\left( x\right) }\right) \leq 0. \] (1) Then \( \left( {S, F}\right) \) is weakly invariant. Let us dispel any doubt as to the meaning of (1): this Hamiltonian inequality says that for every \( \zeta \in {N}_{S}^{P}\left( x\right) \), we have \( h\left( {x,\zeta }\right) \leq 0 \) . Observe that (1) is automatically true if \( {N}_{S}^{P}\left( x\right) \) reduces to \( \{ 0\} \), since \( h\left( {x,0}\right) = 0 \) . Proof. Let us define a function \( {f}_{P} \) as follows: for each \( x \) in \( {\mathbb{R}}^{n} \), choose any \( s = s\left( x\right) \) in \( {\operatorname{proj}}_{S}\left( x\right) \), and let \( v \) in \( F\left( s\right) \) minimize over \( F\left( s\right) \) the function \[ v \mapsto \langle v, x - s\rangle \] We set \( {f}_{P}\left( x\right) = v \) . Note that \( {f}_{P} \) is autonomous; i.e., \( t \) -independent. Since \( x - s \in {N}_{S}^{P}\left( s\right) \), this minimum is nonpositive by (1); i.e., \( \left\langle {{f}_{P}\left( x\right), x - s}\right\rangle \leq 0 \) , which is the main hypothesis of Proposition 2.1. If \( {s}_{0} \) is any given point in \( S \), then \[ \begin{Vmatrix}{{f}_{P}\left( x\right) }\end{Vmatrix} = \parallel v\parallel \leq \gamma \parallel s\parallel + c\text{(since}F\text{has linear growth)} \] \[ \leq \gamma \parallel s - x\parallel + \gamma \parallel x\parallel + c \] \[ = \gamma {d}_{S}\left( x\right) + \gamma \parallel x\parallel + c\left( {\text{since}s \in {\operatorname{proj}}_{S}\left( x\right) }\right) \] \[ \leq \gamma \begin{Vmatrix}{x - {s}_{0}}\end{Vmatrix} + \gamma \parallel x\parallel + c \] \[ \leq {2\gamma }\parallel x\parallel + \begin{Vmatrix}{\gamma {s}_{0}}\end{Vmatrix} + c \] and so \( {f}_{P} \) satisfies the linear growth condition. Now set \( \left\lbrack {a, b}\right\rbrack = \left\lbrack {0,1}\right\rbrack \) and apply Proposition 2.1 for any \( {x}_{0} \in S \) . We conclude that the resulting Euler solutions to \( \dot{x} = {f}_{P}\left( x\right), x\left( 0\right) = {x}_{0} \) on \( \left\lbrack {0,1}\right\rbrack \) necessarily lie in \( S \) . We can extend \( x \) to \( \lbrack 0,\infty ) \) in the evident way by considering next the interval \( \left\lbrack {1,2}\right\rbrack \), etc. The proof will be complete if we can show that \( x \) is a trajectory for \( F \) . (Note that \( f \) is not a selection for \( F \), so Corollary 1.12 is not available.) Let us define another multifunction \( {F}_{S} \) as follows: \[ {F}_{S}\left( x\right) \mathrel{\text{:=}} \operatorname{co}\left\{ {F\left( s\right) : s \in {\operatorname{proj}}_{S}\left( x\right) }\right\} . \] We ask the reader to prove some facts concerning \( {F}_{S} \) . 2.5. Exercise. \( {F}_{S} \) satisfies the Standing Hypotheses, and \( {F}_{S}\left( x\right) = \) \( F\left( x\right) \) if \( x \in S \) . Returning to the proof, observe that by construction, \( {f}_{P} \) is a selection for the multifunction \( {F}_{S} \) . Therefore, by Corollary 1.12, the arc \( x \) defined above is a trajectory for \( {F}_{S} \) ; i.e., \( \dot{x}\left( t\right) \in {F}_{S}\left( {x\left( t\right) }\right) \) a.e. on \( \left\lbrack {0,1}\right\rbrack \) . But since \( x\left( t\right) \in S \) on \( \left\lbrack {0,1}\right\rbrack \), and since \( F = {F}_{S} \) on \( S \), it follows that \( x \) is a trajectory for \( F \) . Note that the proof, a paradigm for other more complicated ones later, is constructive in nature, and actually produces more than is asserted by the theorem, as exemplified by the next corollary. 2.6. Corollary. There is an autonomous function \( {f}_{P} \) with linear growth such that, for any Euler arc \( x \) for \( {f}_{P} \), we have \[ {d}_{S}\left( {x\left( t\right) }\right) \leq {d}_{S}\left( {x\left( a\right) }\right) \forall t \geq a. \] In particular, \( S \) is invariant under \( {f}_{P} \), in the sense that any Euler arc for \( {f}_{P} \) which begins in \( S \) must remain in \( S \) thereafter. The term feedback is used for any function \( f\left( x\right) \) designed or constructed for the purpose of generating Euler solutions. As attractive as the "proximal aiming" feedback \( {f}_{P} \) may be, it does have some drawbacks. It is discontinuous in general, of course, since the metric projection multifunction \( {\operatorname{proj}}_{S} \) does not generally admit continuous selections. In addition, observe that \( {f}_{P}\left( x\right) \) is completely arbitrary within \( F\left( x\right) \) when \( x \in S \) (for then \( s = x \) ). Finally, note that \( {f}_{P}\left( x\right) \in F\left( x\right) \) for \( x \in S \), but not necessarily otherwise: \( {f}_{P} \) is not a selection of \( F \) . The issue of f
1012_(GTM178)Nonsmooth Analysis and Control Theory
68
eft( {x\left( t\right) }\right) \leq {d}_{S}\left( {x\left( a\right) }\right) \forall t \geq a. \] In particular, \( S \) is invariant under \( {f}_{P} \), in the sense that any Euler arc for \( {f}_{P} \) which begins in \( S \) must remain in \( S \) thereafter. The term feedback is used for any function \( f\left( x\right) \) designed or constructed for the purpose of generating Euler solutions. As attractive as the "proximal aiming" feedback \( {f}_{P} \) may be, it does have some drawbacks. It is discontinuous in general, of course, since the metric projection multifunction \( {\operatorname{proj}}_{S} \) does not generally admit continuous selections. In addition, observe that \( {f}_{P}\left( x\right) \) is completely arbitrary within \( F\left( x\right) \) when \( x \in S \) (for then \( s = x \) ). Finally, note that \( {f}_{P}\left( x\right) \in F\left( x\right) \) for \( x \in S \), but not necessarily otherwise: \( {f}_{P} \) is not a selection of \( F \) . The issue of finding feedback selections that lead to invariance will be explored later. First we look at an example of the construction above. 2.7. Exercise. For \( n = 2 \), let \( S = \{ \left( {x, y}\right) : \max \{ x, y\} \geq 0\} \) (the closed complement of the third quadrant), and let \[ F\left( {x, y}\right) = \{ \left( {\left| y\right| - u - 1/2, u}\right) : - 1 \leq u \leq 1\} . \] (a) Show that \( \left( {S, F}\right) \) is weakly invariant via Theorem 2.4. (b) Let the function \( {f}_{P} \) constructed above be expressed in the form \( {f}_{P}\left( {x, y}\right) = \left( {\left| y\right| - u\left( {x, y}\right) - 1/2, u\left( {x, y}\right) }\right) \) . Show that \( u\left( {x, y}\right) = 1 \) for \( x < y < 0, - 1 \) for \( y < x < 0 \), so that \( {f}_{P} \) is discontinuous. What are the possible values of \( {f}_{P} \) when \( x = y < 0 \) ? When \( \left( {x, y}\right) \in S \) ? (c) Prove that no continuous function \( f \) exists having the global invariance property of \( {f}_{P} \) cited in Corollary 2.6. ## Tangential and Other Conditions for Weak Invariance It turns out that weak invariance can be characterized in a number of different ways, one of which involves a multifunction that plays an important role in control theory. 2.8. Definition. The attainable set \( \mathcal{A}\left( {{x}_{0};t}\right) \) for \( t \geq 0 \) is the set of all points of the form \( x\left( t\right) \), where \( x\left( \cdot \right) \) is any trajectory on \( \left\lbrack {0, t}\right\rbrack \) satisfying \( x\left( 0\right) = {x}_{0} \) . 2.9. Exercise. Prove the following: (a) \( \mathcal{A}\left( {{x}_{0};t}\right) \) is compact and nonempty. (b) \( \left( {S, F}\right) \) is weakly invariant \( \Rightarrow \forall {x}_{0} \in S,\forall t > 0,\mathcal{A}\left( {{x}_{0};t}\right) \cap \) \( S \neq \varnothing \) . (c) Semigroup property. For \( \Delta > 0,\mathcal{A}\left( {{x}_{0};t + \Delta }\right) = \mathcal{A}\left( {\mathcal{A}\left( {{x}_{0};t}\right) ;\Delta }\right) \) . (d) For given \( t > 0 \), the multifunction \( x \mapsto \mathcal{A}\left( {x;t}\right) \) satisfies the Standing Hypotheses, except that \( \mathcal{A}\left( {x;t}\right) \) can fail to be a convex set. 2.10. Theorem. The following are equivalent: (a) \( F\left( x\right) \cap {T}_{S}^{B}\left( x\right) \neq \varnothing \forall x \in S \) . (b) \( F\left( x\right) \cap \operatorname{co}{T}_{S}^{B}\left( x\right) \neq \varnothing \forall x \in S \) . (c) \( h\left( {x,{N}_{S}^{P}\left( x\right) }\right) \leq 0\forall x \in S \) . (d) \( \left( {S, F}\right) \) is weakly invariant. (e) \( \forall {x}_{0} \in S,\forall \varepsilon > 0,\exists \delta \in \left( {0,\varepsilon }\right) \) such that \( \mathcal{A}\left( {{x}_{0};\delta }\right) \cap S \neq \varnothing \) . Proof. Condition (a) evidently implies (b). Recall that (Exercise 2.7.1) \[ \operatorname{co}{T}_{S}^{B}\left( x\right) \subseteq {\left\lbrack {N}_{S}^{P}\left( x\right) \right\rbrack }^{ \circ }, \] from which it follows that (b) implies (c). That (c) implies (d) is Theorem 2.4, and the implication (d) \( \Rightarrow \) (e) was noted in Exercise 2.9. Thus only the implication (e) \( \Rightarrow \) (a) remains to be proven. When (e) holds, for a sequence \( {\delta }_{i} \downarrow 0 \), there exist trajectories \( {x}_{i} \) on \( \left\lbrack {0,{\delta }_{i}}\right\rbrack \) with \( {x}_{i}\left( 0\right) = {x}_{0},{x}_{i}\left( {\delta }_{i}\right) \in S \) . Since the trajectories have a common Lipschitz constant, there exists \( K > 0 \) such that \( \left| {{x}_{i}\left( {\delta }_{i}\right) - {x}_{0}}\right| \leq K{\delta }_{i} \) for all large \( i \) . Then, upon taking a subsequence (we eschew relabeling), we can assume that \[ \frac{{x}_{i}\left( {\delta }_{i}\right) - {x}_{0}}{{\delta }_{i}} \rightarrow v \] for some \( v \) . Note that \( v \in {T}_{S}^{B}\left( x\right) \) by definition of the Bouligand tangent cone. Therefore we need only to show that \( v \) lies in \( F\left( x\right) \) in order to deduce (a). We have \[ {x}_{i}\left( {\delta }_{i}\right) - {x}_{0} = {\int }_{0}^{{\delta }_{i}}{\dot{x}}_{i}\left( t\right) {dt} \] (2) where \( {\dot{x}}_{i}\left( t\right) \in F\left( {{x}_{i}\left( t\right) }\right) ,0 \leq t \leq {\delta }_{i} \) . Now for any given \( \Delta > 0 \), for all \( i \) sufficiently large, the set \( \left\{ {{x}_{i}\left( t\right) : 0 \leq t \leq {\delta }_{i}}\right\} \) lies in \( {x}_{0} + {\Delta B} \) . Further, given any \( \varepsilon > 0 \), taking \( \Delta \) small enough will ensure that \( x \in {x}_{0} + {\Delta B} \) implies \( F\left( x\right) \subset F\left( {x}_{0}\right) + {\varepsilon B} \) . The upshot of all this, in view of (2), is that for all \( i \) large enough we have \[ \frac{{x}_{i}\left( {\delta }_{i}\right) - {x}_{0}}{{\delta }_{i}} \in F\left( {x}_{0}\right) + {\varepsilon B} \] Thus \( v \in F\left( {x}_{0}\right) + \varepsilon \bar{B} \) . Since \( \varepsilon \) is arbitrary, \( v \in F\left( {x}_{0}\right) \) as required. ## 2.11. Exercise. (a) \( \left( {S, F}\right) \) is weakly invariant iff given any \( {x}_{0} \in S \) there exists \( \delta > 0 \) (depending on \( {x}_{0} \) ) and a trajectory \( x \) on \( \left\lbrack {0,\delta }\right\rbrack \) such that \( x\left( 0\right) = {x}_{0}, x\left( t\right) \in S\left( {0 \leq t \leq \delta }\right) \) . (Thus, weak invariance can also be characterized in terms of the local existence of a trajectory.) (b) Let \( x \) be a trajectory for \( F \) defined on \( \lbrack a,\infty ) \) . Then the set \[ {S}_{x} \mathrel{\text{:=}} \{ x\left( t\right) : t \geq a\} \] has the property that \( \left( {{S}_{x}, F}\right) \) is weakly invariant. Provide an example in \( {\mathbb{R}}^{2} \) where \( {S}_{x} \) is not closed. (c) The system \( \left( {S, F}\right) \) is termed weakly preinvariant if, given any \( {x}_{0} \in S \), there exists a trajectory \( x \) for \( F \) defined on \( ( - \infty ,0\rbrack \) such that \( x\left( 0\right) = {x}_{0} \) . Show that the system \( \left( {S, F}\right) \) of Exercise 2.7 is weakly invariant but not weakly preinvariant. (d) Prove that \( \left( {S, F}\right) \) is weakly preinvariant iff \( \left( {S, - F}\right) \) is weakly invariant, and that this is in turn equivalent to \[ H\left( {x,{N}_{S}^{P}\left( x\right) }\right) \geq 0\forall x \in S. \] (Hint. Reverse time.) (e) Let \( x \) be a trajectory for \( F \) defined on \( \lbrack a,\infty ) \), and let \[ G \mathrel{\text{:=}} \{ \left( {t, x\left( t\right) }\right) : t \geq a\} . \] Prove that \( G \) is closed, and that \( \left( {G,\{ 1\} \times F}\right) \) is weakly invariant, where \( \{ 1\} \times F \) signifies the multifunction whose value at \( \left( {t, x}\right) \) is the set \( \{ 1\} \times F\left( x\right) \) in \( \mathbb{R} \times {\mathbb{R}}^{n} \) . (f) Show that \( S \) is weakly invariant iff it is the union of trajectories of \( F \) on \( \lbrack 0,\infty ) \) . (g) Show that Theorem 2.10 holds true if the linear growth condition is deleted from the Standing Hypotheses, but \( S \) is assumed to be compact. (Hint. Redefine \( F \) outside a neighborhood of \( S \) so that the new multifunction has uniformly bounded images, as was done in Exercise 1.3(d).) ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_206_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_206_0.jpg) FIGURE 4.1. The set \( S \) of Exercise 2.12. Since \( {T}_{S}^{B}\left( x\right) \) always includes \( {T}_{S}^{C}\left( x\right) \), the tangential or "inwardness" condition that \( F\left( x\right) \cap {T}_{X}^{C}\left( x\right) \neq \varnothing \) for all \( x \in S \) is certainly sufficient for weak invariance of \( \left( {S, F}\right) \) . The next exercise, however, demonstrates that the tangent cone \( {T}_{S}^{C}\left( x\right) \) cannot replace the Bouligand cone in Theorem 2.10, since necessity can fail (possibly only at a single point \( x \) ). 2.12. Exercise. Let \( n = 3 \), and let \( S \mathrel{\text{:=}} \operatorname{cl}\left( {\sqrt{2}B \smallsetminus K}\right) \), where \( K \) (see Figure 4.1) is the cone \[ \left\{ {\left( {{x}_{1},{x}_{2},{x}_{3}}\right) \in {\mathbb{R}}^{3} : \begin{Vmatrix}\left( {{x}_{1},{x}_{2}}\right) \end{Vmatrix} \leq {x}_{3}}\right\} . \] Define \( F \) as follows: let \[ W \mathrel{\text{:=}} \left\{ {\left( {{x}_{1},{x}_{2},1}\right) \in {\mathbb{R}}^{3} : \begin{Vmatrix}\left( {{x}_{1},{x}_{2}}\right) \end{Vmatrix} = 1}\right\} , \] \[ Q \mathrel{\text{:=}} \left\{ {\left( {{x}_{1},{x}_{2},1}\right) \in {\mathbb{R}}^{3} : \begin{Vmatrix}\left( {{x}_{1},{x}_{2}}\right) \end{Vmatrix} \leq 1}\right\} \] and set \[ F\left( {{x}_{1},{x}_{2},{x}_{3}}\right) = F\left( x\right) = \left\{ \begin{array}{ll} Q & \text{ if }x \in S \smallsetminus W \\ \operatorname{co}\left\{ {Q,\left( {-{x}_{2},{x}_{1},0}\right) }\right\} & \text{ otherwise. } \end{array}\right. \] Prove that \( F \) sa
1012_(GTM178)Nonsmooth Analysis and Control Theory
69
). 2.12. Exercise. Let \( n = 3 \), and let \( S \mathrel{\text{:=}} \operatorname{cl}\left( {\sqrt{2}B \smallsetminus K}\right) \), where \( K \) (see Figure 4.1) is the cone \[ \left\{ {\left( {{x}_{1},{x}_{2},{x}_{3}}\right) \in {\mathbb{R}}^{3} : \begin{Vmatrix}\left( {{x}_{1},{x}_{2}}\right) \end{Vmatrix} \leq {x}_{3}}\right\} . \] Define \( F \) as follows: let \[ W \mathrel{\text{:=}} \left\{ {\left( {{x}_{1},{x}_{2},1}\right) \in {\mathbb{R}}^{3} : \begin{Vmatrix}\left( {{x}_{1},{x}_{2}}\right) \end{Vmatrix} = 1}\right\} , \] \[ Q \mathrel{\text{:=}} \left\{ {\left( {{x}_{1},{x}_{2},1}\right) \in {\mathbb{R}}^{3} : \begin{Vmatrix}\left( {{x}_{1},{x}_{2}}\right) \end{Vmatrix} \leq 1}\right\} \] and set \[ F\left( {{x}_{1},{x}_{2},{x}_{3}}\right) = F\left( x\right) = \left\{ \begin{array}{ll} Q & \text{ if }x \in S \smallsetminus W \\ \operatorname{co}\left\{ {Q,\left( {-{x}_{2},{x}_{1},0}\right) }\right\} & \text{ otherwise. } \end{array}\right. \] Prove that \( F \) satisfies the Standing Hypotheses, that \[ F\left( x\right) \cap {T}_{S}^{B}\left( x\right) \neq \varnothing \forall x \in S. \] but that \[ F\left( 0\right) \cap {T}_{S}^{C}\left( 0\right) = \varnothing . \] ## 3 Lipschitz Dependence and Strong Invariance We now wish to focus upon designing feedbacks \( f \) that are actual selections for \( F \) . For example, it can be useful to know whether a given trajectory of \( F \) arises as an Euler solution corresponding to some selection. That this is not generally the case under merely the hypotheses invoked thus far is illustrated by the following. 3.1. Exercise. We define \( F \) on \( {\mathbb{R}}^{2} \) as follows: \[ F\left( {x, y}\right) \mathrel{\text{:=}} \left\{ \begin{array}{ll} \{ \left( {1, e}\right) \} & \text{ if }y > {e}^{x}, \\ \{ \left( {1, r}\right) : 1 \leq r \leq e\} & \text{ if }y = {e}^{x}, \\ \{ \left( {1,1}\right) \} & \text{ if }y < {e}^{x}. \end{array}\right. \] (a) Confirm that \( F \) satisfies the Standing Hypotheses, and that \( \left( {\bar{x},\bar{y}}\right) \left( t\right) = \left( {t,{e}^{t}}\right) \) is a trajectory on \( \left\lbrack {0,1}\right\rbrack \) . Show that there is no selection \( f \) for \( F \) (autonomous or not) admitting \( \left( {\bar{x},\bar{y}}\right) \) as an Euler solution of \( \left( {\dot{x},\dot{y}}\right) = f\left( {t, x, y}\right) ,\left( {x\left( 0\right), y\left( 0\right) }\right) = \left( {0,1}\right) \) . (b) The set \( \left\{ {\left( {t,{e}^{t}}\right) : t \geq 0}\right\} \) is weakly invariant with respect to \( F \), by part (b) of Exercise 2.11. Deduce from this that there exist weakly invariant systems which fail to admit a feedback selection \( f \) whose Euler arcs leave \( S \) invariant. A better behavior on the part of \( F \) that will preclude such pathology is provided by a multivalued version of local Lipschitz behavior: 3.2. Definition. \( F \) is said to be locally Lipschitz provided that every point \( x \) admits a neighborhood \( U = U\left( x\right) \) and a positive constant \( K = K\left( x\right) \) such that \[ {x}_{1},{x}_{2} \in U \Rightarrow F\left( {x}_{2}\right) \subseteq F\left( {x}_{1}\right) + K\begin{Vmatrix}{{x}_{1} - {x}_{2}}\end{Vmatrix}\bar{B}. \] (1) We then say that \( F \) is Lipschitz of rank \( K \) on the set \( U \) . 3.3. Exercise. Let \( F \) be locally Lipschitz. (a) Let \( C \) be any bounded subset of \( {\mathbb{R}}^{n} \) . Show that there exists \( K \) such that (1) holds on \( C \) ; thus \( F \) is "Lipschitz on bounded sets.” (b) Fix \( {v}_{0} \in {\mathbb{R}}^{n} \), and let \( f\left( x\right) \) be the point \( v \) in \( F\left( x\right) \) closest to \( {v}_{0} \) . Prove that \( f \) is a continuous selection for \( F \) . (In general, \( f \) need not be locally Lipschitz.) (c) Prove that \( F \) is locally Lipschitz iff \( x \mapsto h\left( {x, p}\right) \) is locally Lipschitz for each \( p \) . We now confirm that when \( F \) is locally Lipschitz, a weakly invariant system \( \left( {S, F}\right) \) admits a feedback selection whose Euler arcs all leave \( S \) invariant. 3.4. Theorem. Let \( \left( {S, F}\right) \) be weakly invariant, where \( F \) is locally Lipschitz. Then there exists a feedback selection \( {g}_{P} \) for \( F \) under which \( S \) is invariant; that is, such that for any Euler arc \( x \) for \( {g}_{P} \) on \( \left\lbrack {a, b}\right\rbrack \) having \( x\left( a\right) \in S \), we have \( x\left( t\right) \in S\forall t \in \left\lbrack {a, b}\right\rbrack \) . Proof. Let \( {f}_{P} \) be defined precisely as in the proof of Theorem 2.4. Then \( {f}_{P}\left( x\right) \) lies not necessarily in \( F\left( x\right) \), but in \( F\left( s\right) \), where \( s = s\left( x\right) \in {\operatorname{proj}}_{S}\left( x\right) \) . We define \( {g}_{P}\left( x\right) \) to be the point in \( F\left( x\right) \) closest to \( {f}_{P}\left( x\right) \), so that \( {g}_{P} \) is a selection for \( F \) . Now let \( {x}_{0} \in S \) and \( b > a \) be given. We will complete the proof by showing that any Euler solution \( \bar{x} \) on \( \left\lbrack {a, b}\right\rbrack \) from \( {x}_{0} \) generated by \( {g}_{P} \) is such that \( \bar{x}\left( t\right) \in S\forall t \in \left\lbrack {a, b}\right\rbrack \) . (We already know that \( \bar{x} \) is a trajectory from Corollary 1.12.) It follows from Exercise 1.9 that there is some a priori bound for \( \bar{x}\left( t\right) \) on \( \left\lbrack {a, b}\right\rbrack \), let us say \( \begin{Vmatrix}{\bar{x}\left( t\right) - {x}_{0}}\end{Vmatrix} < M \) . Let \( K \) be a Lipschitz constant for \( F \) on the set \( {x}_{0} + {2MB} \) . Note now that if \( \begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix} < M \), then \[ \begin{Vmatrix}{s - {x}_{0}}\end{Vmatrix} \leq \parallel s - x\parallel + \begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix} = {d}_{S}\left( x\right) + \begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix} \] \[ \leq 2\begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix} < {2M}\text{.} \] This allows us to estimate as follows: \[ \langle {g}_{P}\left( x\right), x - s\rangle = \langle {f}_{P}\left( x\right), x - s\rangle + \langle {g}_{P}\left( x\right) - {f}_{P}\left( x\right), x - s\rangle \] \[ \leq \begin{Vmatrix}{{g}_{P}\left( x\right) - {f}_{P}\left( x\right) }\end{Vmatrix}\parallel x - s\parallel \] \[ \leq K\parallel x - s{\parallel }^{2} = K{d}_{S}^{2}\left( x\right) . \] This shows that the extension of Proposition 2.1 provided by Exercise 2.2 is applicable, with \( \theta \left( {t, x}\right) \mathrel{\text{:=}} K{d}_{S}\left( x\right) \) . Letting \( \bar{x} \) be an Euler solution corresponding to \( {g}_{P} \), we conclude that \[ \frac{d}{dt}{d}_{S}\left( {\bar{x}\left( t\right) }\right) \leq K{d}_{S}\left( {\bar{x}\left( t\right) }\right) \text{ a.e. in }\left\lbrack {a, b}\right\rbrack . \] The Gronwall inequality then implies that \[ {d}_{S}\left( {\bar{x}\left( t\right) }\right) \leq {d}_{S}\left( {\bar{x}\left( a\right) }\right) {e}^{K\left( {t - a}\right) } = 0, \] since \( \bar{x}\left( a\right) = {x}_{0} \in S \) . Thus \( \bar{x}\left( t\right) \in S, a \leq t \leq b \) . The proof yields the following estimate for solutions not necessarily beginning in \( S \) : 3.5. Corollary. Let \( K \) be a Lipschitz constant for \( F \) on the set \( {x}_{0} + {2MB} \) , and let \( b > a \) satisfy \[ \left( {b - a}\right) {e}^{\gamma \left( {b - a}\right) }\left( {c + \gamma \begin{Vmatrix}{x}_{0}\end{Vmatrix}}\right) < M. \] Then any Euler solution on \( \left\lbrack {a, b}\right\rbrack \) of \( \dot{x} = {g}_{P}\left( x\right), x\left( a\right) = {x}_{0} \) is a trajectory satisfying \[ {d}_{S}\left( {x\left( t\right) }\right) \leq {d}_{S}\left( {x}_{0}\right) {e}^{K\left( {t - a}\right) },\;a \leq t \leq b. \] Proof. The inequalities in the proof hold for \( \begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix} < M \), for the given \( K \) ; on the other hand, the solutions \( x\left( t\right) \) continue to satisfy \( \begin{Vmatrix}{x\left( t\right) - {x}_{0}}\end{Vmatrix} < M \) for \( t \in \left\lbrack {a, b}\right\rbrack \) when \( b - a \) is small enough (as stated), according to Theorem 1.7. Thus, at least on \( \left\lbrack {a, b}\right\rbrack \), the upper bound on \( {d}_{S}\left( {x\left( t\right) }\right) \) ensues. 3.6. Exercise. Observe that the multifunction \( F \) of Exercise 2.7 is locally Lipschitz. What is the feedback selection \( {g}_{P} \) of Theorem 3.4 in this case? The following consequence affirms that for Lipschitz \( F \), any trajectory can be generated by a feedback selection, which resolves an earlier question. 3.7. Corollary. When \( F \) is locally Lipschitz, a given arc \( \bar{x} \) on \( \left\lbrack {a, b}\right\rbrack \) is a trajectory for \( F \) iff there exists a (not necessarily autonomous) feedback selection \( f \) for \( F \) such that \( \bar{x} \) is an Euler solution on \( \left\lbrack {a, b}\right\rbrack \) for the initial-value problem \( \dot{x} = f\left( {t, x}\right), x\left( a\right) = \bar{x}\left( a\right) \) . In fact, given a trajectory \( \bar{x} \), there is a feedback selection for \( F \) for which \( \bar{x} \) is the unique Euler solution of the associated initial-value problem. Proof. That selections give rise to trajectories was noted in Corollary 1.12, so it remains to exhibit, for a given trajectory \( \bar{x} \), a feedback selection generating it. The set \( G \) of Exercise 2.11(e) is weakly invariant under the trajectories of \( \{ 1\} \times F \), allowing us to invoke Theorem 3.4. This provides a selection for \( \{ 1\} \times F \), necessarily of the form \( \left( {1, f\left( {t, x}\right) }\right) \), where \( f \) is a selection for \( F \), and having the property that any Euler solution \( x \) of the initial-value problem \( \dot{x} = f\left( {t, x}\right), x\left( a\right) = \bar{x}\left( a\right) \) is such that \( \left( {t, x\left( t\right) }\right) \in G \) on \
1012_(GTM178)Nonsmooth Analysis and Control Theory
70
tial-value problem \( \dot{x} = f\left( {t, x}\right), x\left( a\right) = \bar{x}\left( a\right) \) . In fact, given a trajectory \( \bar{x} \), there is a feedback selection for \( F \) for which \( \bar{x} \) is the unique Euler solution of the associated initial-value problem. Proof. That selections give rise to trajectories was noted in Corollary 1.12, so it remains to exhibit, for a given trajectory \( \bar{x} \), a feedback selection generating it. The set \( G \) of Exercise 2.11(e) is weakly invariant under the trajectories of \( \{ 1\} \times F \), allowing us to invoke Theorem 3.4. This provides a selection for \( \{ 1\} \times F \), necessarily of the form \( \left( {1, f\left( {t, x}\right) }\right) \), where \( f \) is a selection for \( F \), and having the property that any Euler solution \( x \) of the initial-value problem \( \dot{x} = f\left( {t, x}\right), x\left( a\right) = \bar{x}\left( a\right) \) is such that \( \left( {t, x\left( t\right) }\right) \in G \) on \( \left\lbrack {a, b}\right\rbrack \) . But then \( x\left( t\right) = \bar{x}\left( t\right) \) for \( t \in \left\lbrack {a, b}\right\rbrack \), proving that \( \bar{x} \) is the (unique) Euler solution. ## Strong Invariance The system \( \left( {S, F}\right) \) is said to be strongly invariant if every trajectory \( x \) on \( \lbrack 0,\infty ) \) for which \( x\left( 0\right) \in S \) is such that \( x\left( t\right) \in S \) for all \( t \geq 0 \) . Remaining still in the case in which \( F \) is autonomous, let us characterize strong invariance in terms resembling those of Theorem 2.10, but with a Lipschitz hypothesis imposed upon the multifunction \( F \) . Note that now \( {T}_{S}^{C} \), the tangent cone of \( \$ {2.5} \), joins the cast by playing a role in one of the criteria for strong invariance; in this regard recall Exercise 2.12, which showed that \( {T}_{S}^{C} \) could not be included in the characterizations of weak invariance. Note that the upper Hamiltonian \( H \) figures in (d) below, so that (d) is a strong counterpart of the weak proximal normal condition that figured in Theorem 2.10. 3.8. Theorem. Let \( F \) be locally Lipschitz. Then the following are equivalent: (a) \( F\left( x\right) \subseteq {T}_{S}^{C}\left( x\right) \forall x \in S \) . (b) \( F\left( x\right) \subseteq {T}_{S}^{B}\left( x\right) \forall x \in S \) . (c) \( F\left( x\right) \subseteq \operatorname{co}{T}_{S}^{B}\left( x\right) \forall x \in S \) . (d) \( H\left( {x,{N}_{S}^{P}\left( x\right) }\right) \leq 0\forall x \in S \) . (e) \( \left( {S, F}\right) \) is strongly invariant. (f) \( \forall {x}_{0} \in S,\exists \varepsilon > 0 \ni \mathcal{A}\left( {{x}_{0};t}\right) \subseteq S\forall t \in \left\lbrack {0,\varepsilon }\right\rbrack \) . Proof. That (a) \( \Rightarrow \) (b) \( \Rightarrow \) (c) is a tautology, since \( {T}_{S}^{C}\left( x\right) \subseteq {T}_{S}^{B}\left( x\right) \) . Since every element of \( \operatorname{co}{T}_{S}^{B}\left( x\right) \) belongs to \( {N}_{S}^{P}{\left( x\right) }^{ \circ } \), we know (c) \( \Rightarrow \) (d). Let us now address the implication \( \left( \mathrm{d}\right) \Rightarrow \left( \mathrm{e}\right) \) . Let \( \bar{x} \) be any trajectory for \( F \) on \( \left\lbrack {a, b}\right\rbrack \), with \( \bar{x}\left( a\right) \in S \) . According to Corollary 3.7, there exists a feedback selection \( f \) such that \( \bar{x} \) is an Euler solution of the initial-value problem \( \dot{x} = f\left( {t, x}\right), x\left( a\right) = \bar{x}\left( a\right) = : {x}_{0} \) . Let \( M > 0 \) be such that all Euler solutions \( x \) of this initial-value problem above satisfy \( \begin{Vmatrix}{x\left( t\right) - {x}_{0}}\end{Vmatrix} < M, a \leq t \leq b \) . If \( x \) lies in \( {x}_{0} + {MB} \) and \( s \in {\operatorname{proj}}_{S}\left( x\right) \), then \[ \begin{Vmatrix}{s - {x}_{0}}\end{Vmatrix} \leq \parallel s - x\parallel + \begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix} \leq 2\begin{Vmatrix}{x - {x}_{0}}\end{Vmatrix}, \] so that \( s \in {x}_{0} + {2MB} \) . Now let \( K \) be a Lipschitz constant for \( F \) on \( {x}_{0} + {2MB} \), and consider any \( x \in {x}_{0} + {MB} \) and \( s \in {\operatorname{proj}}_{S}\left( x\right) \) . Then \( x - s \in {N}_{S}^{P}\left( s\right) \) . Since \( f\left( {t, x}\right) \in F\left( x\right) \) , there exists \( v \in F\left( s\right) \) such that \( \parallel v - f\left( {t, x}\right) \parallel \leq K\parallel s - x\parallel = K{d}_{S}\left( x\right) \) . Further, by condition (d), we have \( \langle v, x - s\rangle \leq 0 \) . We deduce \[ \langle f\left( {t, x}\right), x - s\rangle \leq K{d}_{S}{\left( x\right) }^{2}. \] Thus a special case of Exercise 2.2 holds, for \( \theta \left( {t, x}\right) \mathrel{\text{:=}} K{d}_{S}\left( x\right) \), and we obtain \[ \frac{d}{dt}{d}_{S}\left( {\bar{x}\left( t\right) }\right) \leq K{d}_{S}\left( {\bar{x}\left( t\right) }\right) ,\;a \leq t \leq b,\;{d}_{S}\left( {\bar{x}\left( 0\right) }\right) = 0, \] which implies that \( {d}_{S}\left( {\bar{x}\left( t\right) }\right) = 0 \) in \( \left\lbrack {a, b}\right\rbrack \) by the now familiar Gronwall inequality. Thus \( \bar{x}\left( t\right) \in S\forall t \in \left\lbrack {a, b}\right\rbrack \), and (e) holds. That (e) and (f) are equivalent is easy to see; let us now show (e) \( \Rightarrow \) (d). Consider any \( \widetilde{x} \in S \), let any \( \widetilde{v} \) in \( F\left( \widetilde{x}\right) \) be given, and set \( \widetilde{F}\left( x\right) = \{ \widetilde{f}\left( x\right) \} \) , where \( \widetilde{f}\left( x\right) \) is the closest point in \( F\left( x\right) \) to \( \widetilde{v} \) . Note that \( \widetilde{f}\left( \widetilde{x}\right) = \widetilde{v} \), and that \( \widetilde{f} \) is a continuous selection of \( F \) (Exercise 3.3(c)). Thus \( \widetilde{F} \) satisfies the Standing Hypotheses, and clearly \( \left( {S,\widetilde{F}}\right) \) is strongly, and hence weakly, invariant. By Theorem 2.10, we must have \[ \widetilde{h}\left( {\widetilde{x},{N}_{S}^{P}\left( \widetilde{x}\right) }\right) \leq 0 \] here, \( \widetilde{h} \) of course denotes the lower Hamiltonian associated with \( \widetilde{F} \) . This is the same as asserting that \( \langle \widetilde{v},\zeta \rangle \) is nonpositive for any \( \zeta \in {N}_{S}^{P}\left( \widetilde{x}\right) \) . Since \( \widetilde{v} \) is arbitrary in \( F\left( \widetilde{x}\right) \) ,(d) follows. To complete the proof of the theorem, it suffices to prove that (d) \( \Rightarrow \) (a). Let \( \widetilde{v} \) be any element of \( F\left( \widetilde{x}\right) \) ; we need to show that \( \widetilde{v} \) belongs to \( {N}_{S}^{L}{\left( \widetilde{x}\right) }^{ \circ } = \) \( {T}_{S}^{C}\left( \widetilde{x}\right) \) . Now any element \( \zeta \) of \( {N}_{S}^{L}\left( \widetilde{x}\right) \) is of the form \( \zeta = \mathop{\lim }\limits_{i}{\zeta }_{i} \), where \( {\zeta }_{i} \in {N}_{S}^{P}\left( {x}_{i}\right) \) and \( {x}_{i} \rightarrow \widetilde{x} \) . For each \( i \), there exists \( {v}_{i} \in F\left( {x}_{i}\right) \) such that \[ \begin{Vmatrix}{{v}_{i} - \widetilde{v}}\end{Vmatrix} \leq K\begin{Vmatrix}{{x}_{i} - \widetilde{x}}\end{Vmatrix} \] and (by (d)) we have \( \left\langle {{\zeta }_{i},{v}_{i}}\right\rangle \leq 0 \) . We derive \( \langle \zeta ,\widetilde{v}\rangle \leq 0 \) as required. 3.9. Exercise. (a) Show that the local Lipschitz assumption on \( F \) cannot be dispensed with in the preceding theorem, by taking \( F\left( {t, x}\right) = \) \( \{ f\left( x\right) \} \), and \( S = \{ 0\} \), where \( f \) is the function of Exercise 1.10(b). (b) Show that the system of Exercise 2.7 is weakly but not strongly invariant. ## The Nonautonomous Case The invariance results obtained in the preceding sections can all be extended to the case in which \( F \) depends on \( t \) as well as on \( x \) (the nonautonomous case), and there is an easy route for getting the nonautonomous results from the autonomous versions. This technique is called state augmentation; here, it consists of viewing \( t \) as simply a component of the state, one whose derivative is always 1 . To be more precise, we will think of \( t \) as being the zeroth coordinate of \( x \), and we implement this notationally through the convention wherein \( \bar{x} \) denotes a vector \( \left( {{x}^{0}, x}\right) \) in \( \mathbb{R} \times {\mathbb{R}}^{n} \) . We proceed to define an augmented multifunction \( \bar{F} \) which satisfies the Standing Hypotheses whenever \( F \) does: \[ \bar{F}\left( \bar{x}\right) = \bar{F}\left( {{x}^{0}, x}\right) = \{ 1\} \times F\left( {{x}^{0}, x}\right) . \] Then if \( \bar{x}\left( \cdot \right) = \left( {{x}^{0}\left( \cdot \right), x\left( \cdot \right) }\right) \) is a trajectory for \( \bar{F} \), with \( \bar{x}\left( a\right) = {\bar{x}}_{0} = \left( {{x}_{0}^{0},{x}_{0}}\right) \) , it follows that \( x \) is a trajectory for \( F \) with \( x\left( a\right) = {x}_{0} \), and that \( {x}^{0}\left( t\right) = \) \( {x}_{0}^{0} + t - a \) . Conversely, if \( x \) is a trajectory for \( F \), we augment it to a trajectory \( \bar{x} \) for \( \bar{F} \) by setting \( {x}^{0}\left( t\right) = {x}_{0}^{0} + t - a \) (for any choice of \( {x}_{0}^{0} \) ). As before, let \( S \) be a given closed subset of \( {\mathbb{R}}^{n} \), and now let \( F\left( {t, x}\right) \) possibly depend on \( t \) . We extend Definition 2.3 as follows: the system \( \left( {S, F}\right) \) is called weakly invariant provided that for all \( \left( {{t}_{0},{x}_{0}}\right) \) with \( {x}_{0} \in S \), there exists a trajectory \( x \) of \( F \) on \( \left\lbrack {{t}_{0},\infty }\right) \) satisfying \( x\left( {t}_{0}\right) = {x}_{0}, x\left( t\right) \in S\forall t \geq {t}_{0} \) . 3.10. Exercise. For the nonautonomous case described above, show that \( \left( {S, F}\right) \) is weakly invarian
1012_(GTM178)Nonsmooth Analysis and Control Theory
71
t \( x \) is a trajectory for \( F \) with \( x\left( a\right) = {x}_{0} \), and that \( {x}^{0}\left( t\right) = \) \( {x}_{0}^{0} + t - a \) . Conversely, if \( x \) is a trajectory for \( F \), we augment it to a trajectory \( \bar{x} \) for \( \bar{F} \) by setting \( {x}^{0}\left( t\right) = {x}_{0}^{0} + t - a \) (for any choice of \( {x}_{0}^{0} \) ). As before, let \( S \) be a given closed subset of \( {\mathbb{R}}^{n} \), and now let \( F\left( {t, x}\right) \) possibly depend on \( t \) . We extend Definition 2.3 as follows: the system \( \left( {S, F}\right) \) is called weakly invariant provided that for all \( \left( {{t}_{0},{x}_{0}}\right) \) with \( {x}_{0} \in S \), there exists a trajectory \( x \) of \( F \) on \( \left\lbrack {{t}_{0},\infty }\right) \) satisfying \( x\left( {t}_{0}\right) = {x}_{0}, x\left( t\right) \in S\forall t \geq {t}_{0} \) . 3.10. Exercise. For the nonautonomous case described above, show that \( \left( {S, F}\right) \) is weakly invariant iff \[ h\left( {t, x,{N}_{S}^{P}\left( x\right) }\right) \leq 0\forall t \in \mathbb{R},\forall x \in S. \] When \( F \) is locally Lipschitz, show that we obtain the corresponding characterization of strong invariance. (Hint. Augment the state, and consider \( \bar{S} \mathrel{\text{:=}} \mathbb{R} \times S \) .) ## Dependence on Initial Conditions In the general case in which \( F\left( {t, x}\right) \) is nonautonomous, which remains our interest here, the attainable set depends of course on the initial-value of \( t \) as well. Thus, \( \mathcal{A}\left( {{t}_{0},{x}_{0};T}\right) \), for \( T \geq {t}_{0} \), is defined as the set of all points of the form \( x\left( T\right) \), where \( x \) is a trajectory for \( F \) on \( \left\lbrack {{t}_{0}, T}\right\rbrack \) satisfying \( x\left( {t}_{0}\right) = {x}_{0} \) . Classically, the theorem by which the solution of a differential equation depends regularly on initial conditions is a well-known and useful result. Here is its counterpart for differential inclusions: 3.11. Theorem. Let \( F\left( {t, x}\right) \) be locally Lipschitz in \( \left( {t, x}\right) \) . Then for any fixed \( T \in \mathbb{R} \), the multifunction \( \left( {{t}_{0},{x}_{0}}\right) \mapsto \mathcal{A}\left( {{t}_{0},{x}_{0};T}\right) \) is locally Lipschitz on \( ( - \infty, T\rbrack \times {\mathbb{R}}^{n} \) . Proof. Fix \( a < T \) and \( y \in {\mathbb{R}}^{n} \) . We will exhibit a constant \( \rho \) with the property that for any \( \left( {{t}_{0},{x}_{0}}\right) \in \lbrack a, T) \times B\left( {y;1}\right) \), for any trajectory \( z \) on \( \left\lbrack {{t}_{0}, T}\right\rbrack \) having \( z\left( {t}_{0}\right) = {x}_{0} \), and for any other \( \left( {\tau ,\alpha }\right) \in \lbrack a, T) \times B\left( {y;1}\right) \), there exists a trajectory \( x \) on \( \left\lbrack {\tau, T}\right\rbrack \) with \( x\left( \tau \right) = \alpha \) such that \[ \parallel x\left( T\right) - z\left( T\right) \parallel \leq \rho \begin{Vmatrix}\left( {\tau - {t}_{0},\alpha - {x}_{0}}\right) \end{Vmatrix}. \] This will establish that the mapping \( \left( {t, x}\right) \mapsto \mathcal{A}\left( {t, x;T}\right) \) is locally Lipschitz on \( \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \) ; the extension to \( ( - \infty, T\rbrack \) will be treated subsequently. We augment the state as described above, setting \[ \bar{z}\left( t\right) = \left( {t, z\left( t\right) }\right) ,\;{\bar{z}}_{0} = \left( {{t}_{0},{x}_{0}}\right) ,\;G = \{ \left( {t, z\left( t\right) }\right) : a \leq t < \infty \} , \] where we have extended \( z \) to the interval \( \lbrack a,\infty ) \) . Then \( \bar{z} \) is a trajectory on \( \left\lbrack {a, T}\right\rbrack \) for the multifunction \( \bar{F} \mathrel{\text{:=}} \{ 1\} \times F \), so that \( \left( {G,\bar{F}}\right) \) is weakly invariant. This allows us to invoke Corollary 3.5 to deduce the existence of an arc \( \bar{x}\left( t\right) = \left( {t, x\left( t\right) }\right) \) on \( \left\lbrack {\tau, T}\right\rbrack \) such that \( x\left( \tau \right) = \alpha, x \) is a trajectory of \( F \) , and \[ {d}_{G}\left( {t, x\left( t\right) }\right) \leq {e}^{K\left( {T - \tau }\right) }{d}_{G}\left( {\tau ,\alpha }\right) \] where \( K \) is a Lipschitz constant for \( F \) on an appropriately large ball. Then \( x \) is a trajectory for \( F \) on \( \left\lbrack {\tau, T}\right\rbrack \) and we have \[ {d}_{G}\left( {T, x\left( T\right) }\right) \leq {e}^{K\left( {T - a}\right) }\begin{Vmatrix}\left( {\tau - {t}_{0},\alpha - {x}_{0}}\right) \end{Vmatrix}, \] since \( \left( {{t}_{0},{x}_{0}}\right) \in G \) . Accordingly, for some point \( \left( {{t}^{\prime }, z\left( {t}^{\prime }\right) }\right) \in G \) we have \[ \begin{Vmatrix}\left( {T - {t}^{\prime }, x\left( T\right) - z\left( {t}^{\prime }\right) }\right) \end{Vmatrix} \leq {e}^{K\left( {T - a}\right) }\begin{Vmatrix}\left( {\tau - {t}_{0},\alpha - {x}_{0}}\right) \end{Vmatrix}. \] Let \( k \geq 1 \) be a common Lipschitz constant for all trajectories of \( F \) on \( \left\lbrack {a, T}\right\rbrack \) with initial-value in \( B\left( {y;1}\right) \) . Then \[ \parallel z\left( T\right) - x\left( T\right) \parallel \leq \begin{Vmatrix}{z\left( T\right) - z\left( {t}^{\prime }\right) }\end{Vmatrix} + \begin{Vmatrix}{z\left( {t}^{\prime }\right) - x\left( T\right) }\end{Vmatrix} \] \[ \leq k\left| {T - {t}^{\prime }}\right| + k\begin{Vmatrix}{z\left( {t}^{\prime }\right) - x\left( T\right) }\end{Vmatrix} \leq {2k}\begin{Vmatrix}\left( {T - {t}^{\prime }, x\left( T\right) - z\left( {t}^{\prime }\right) }\right) \end{Vmatrix} \] \[ \leq {2k}{e}^{K\left( {T - a}\right) }\begin{Vmatrix}\left( {\tau - {t}_{0},\alpha - {x}_{0}}\right) \end{Vmatrix}. \] This reveals the required constant \( \rho \mathrel{\text{:=}} {2k}{e}^{K\left( {T - a}\right) } \) . To complete the proof, observe that \( \mathcal{A}\left( {T,{x}_{0};T}\right) = \left\{ {x}_{0}\right\} \), so it suffices to show that \[ \mathcal{A}\left( {\tau ,\alpha ;T}\right) \subset \left\{ {x}_{0}\right\} + \rho \begin{Vmatrix}\left( {T - \tau ,{x}_{0} - \alpha }\right) \end{Vmatrix}\bar{B} \] But if \( x \) is any trajectory on \( \left\lbrack {\tau, T}\right\rbrack \) with \( x\left( \tau \right) = \alpha \), then \[ \begin{Vmatrix}{x\left( T\right) - {x}_{0}}\end{Vmatrix} \leq \parallel x\left( T\right) - \alpha \parallel + \begin{Vmatrix}{\alpha - {x}_{0}}\end{Vmatrix} \] \[ = \parallel x\left( T\right) - x\left( \tau \right) \parallel + \begin{Vmatrix}{\alpha - {x}_{0}}\end{Vmatrix} \leq k\left| {T - \tau }\right| + k\begin{Vmatrix}{\alpha - {x}_{0}}\end{Vmatrix} \] \[ \leq {2k}\begin{Vmatrix}\left( {T - \tau ,{x}_{0} - \alpha }\right) \end{Vmatrix} \leq \rho \begin{Vmatrix}\left( {T - \tau ,{x}_{0} - \alpha }\right) \end{Vmatrix}, \] which completes the proof. The following exercise provides useful facts regarding an optimal control problem to be studied later. 3.12. Exercise. Let \( \ell : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be continuous. For \( \tau \leq T \) and for any \( \alpha \in {\mathbb{R}}^{n} \), set \[ V\left( {\tau ,\alpha }\right) \mathrel{\text{:=}} \inf \{ \ell \left( {x\left( T\right) }\right) : x\text{ is a trajectory of }F\text{ on }\left\lbrack {\tau, T}\right\rbrack \text{ with }x\left( \tau \right) = \alpha \} . \] (a) Prove that the infimum defining \( V\left( {\tau ,\alpha }\right) \) is attained. (b) Now let \( F \) be locally Lipschitz. Prove that \( V \) is continuous on \( ( - \infty, T\rbrack \times {\mathbb{R}}^{n} \), and locally Lipschitz if \( \ell \) is locally Lipschitz. ## 4 Equilibria Consider the issue of stabilizing at a given point \( {x}^{ * } \) a system described as the trajectory \( x\left( t\right) \) of an autonomous differential inclusion \( \dot{x} \in F\left( x\right) \) . If the system is already at \( {x}^{ * } \), then it will be possible to remain there iff \( 0 \in F\left( {x}^{ * }\right) \) , a condition which is described by saying that \( {x}^{ * } \) is a zero (or equilibrium, or rest point) of the multifunction \( F \) . Note that this is equivalent to the weak invariance of the system \( \left( {\left\{ {x}^{ * }\right\}, F}\right) \), so that formally, the study of equilibria amounts to that of invariant singleton sets, an observation that motivates the dynamic approach to the issue that we will develop in this section. ## The Classical Case and Brouwer's Theorem At the heart of very many theorems on the existence of equilibria is the celebrated Brouwer Fixed Point Theorem, which can be stated as follows: If \( g \) is a continuous function mapping \( \bar{B} \) to itself, where \( \bar{B} \) is the closed unit ball in \( {\mathbb{R}}^{n} \), then there is a point \( u \) in \( \bar{B} \) such that \( g\left( u\right) = u \) . Let us note the bearing of this theorem upon the differential equation \[ \dot{x}\left( t\right) = f\left( {x\left( t\right) }\right) \] (1) where \( f \) is a Lipschitz mapping from \( {\mathbb{R}}^{n} \) to \( {\mathbb{R}}^{n} \) . Suppose that the solution \( x\left( t\right) \) to (1), for any initial condition \( x\left( 0\right) = {x}_{0} \in \bar{B} \), has the property that \( x\left( t\right) \in \bar{B}\forall t \geq 0 \) . That is, suppose that the system \( \left( {\bar{B}, f}\right) \) is invariant; "weakly" and "strongly" coalesce in this setting. It follows from the Brouwer Fixed Point Theorem that \( f \) must admit a zero in \( \bar{B} \) . This and more is taken up in the following exercise: ## 4.1. Exercise. (a) Prove that in the situation described above there is a point \( {x}^{ * } \) in \( \bar{B} \) such that \( f\left( {x}^{ * }\right) = 0 \) . (Hint. For given \( \tau > 0 \), consider the map \( {f}_{\tau }\left( \alpha \right) \mathrel{\text{:=}} x\left( {\tau ;\alpha }\right) \) ; that is, the value of \( x\left( \tau \right) \), where \( x \) is the (unique) solution
1012_(GTM178)Nonsmooth Analysis and Control Theory
72
t\right) }\right) \] (1) where \( f \) is a Lipschitz mapping from \( {\mathbb{R}}^{n} \) to \( {\mathbb{R}}^{n} \) . Suppose that the solution \( x\left( t\right) \) to (1), for any initial condition \( x\left( 0\right) = {x}_{0} \in \bar{B} \), has the property that \( x\left( t\right) \in \bar{B}\forall t \geq 0 \) . That is, suppose that the system \( \left( {\bar{B}, f}\right) \) is invariant; "weakly" and "strongly" coalesce in this setting. It follows from the Brouwer Fixed Point Theorem that \( f \) must admit a zero in \( \bar{B} \) . This and more is taken up in the following exercise: ## 4.1. Exercise. (a) Prove that in the situation described above there is a point \( {x}^{ * } \) in \( \bar{B} \) such that \( f\left( {x}^{ * }\right) = 0 \) . (Hint. For given \( \tau > 0 \), consider the map \( {f}_{\tau }\left( \alpha \right) \mathrel{\text{:=}} x\left( {\tau ;\alpha }\right) \) ; that is, the value of \( x\left( \tau \right) \), where \( x \) is the (unique) solution of the initial-value problem satisfying the initial condition \( x\left( 0\right) = \alpha \) . Invoke the Brouwer Fixed Point Theorem to deduce that for each \( \tau > 0 \) there exists \( {x}_{\tau } \in \bar{B} \) such that \( x\left( {\tau ;{x}_{\tau }}\right) = {x}_{\tau } \), and then let \( \tau \downarrow 0 \) .) (b) Conversely, given the "zero point theorem" of part (a), deduce from it the Brouwer Fixed Point Theorem. (Hint. Given \( g \), consider \( f\left( x\right) \mathrel{\text{:=}} g\left( x\right) - x \) .) We would now like to consider sets \( S \) other than the ball, in fact sets that are not even convex. It is certainly the case, however, that some topological hypothesis will have to be made. Consider for example the annulus \( S \) in \( {\mathbb{R}}^{2} \) described in polar coordinates \( \left( {r,\theta }\right) \) via \( S \mathrel{\text{:=}} \{ \left( {r,\theta }\right) : 1 \leq r \leq 2\} \), and let \( g\left( {r,\theta }\right) \mathrel{\text{:=}} \left( {r,\theta + {\theta }_{0}}\right) \) for \( 0 < {\theta }_{0} < {2\pi } \) . Then \( g \) is a smooth function mapping \( S \) to itself, yet \( g \) admits no fixed point. One immediate extension of the Brouwer Fixed Point Theorem is obtained by taking \( S \) to be homeomorphic to a closed unit ball. This means that \( S = h\left( \bar{B}\right) \) where \( \bar{B} \) is the closed unit ball in \( {\mathbb{R}}^{k}, h : \bar{B} \rightarrow S \) is continuous, and where the inverse \( {h}^{-1} : S \rightarrow \bar{B} \) exists and is continuous. 4.2. Exercise. (a) Let \( g : S \rightarrow S \) be continuous, and let \( S \) be homeomorphic to a closed unit ball. Prove that \( g \) has a fixed point in \( S \) . (b) Show that part (a) of Exercise 4.1 holds true if the unit ball is replaced by a set which is homeomorphic to the ball. ## A Conjecture in the Multivalued Case The next step is to replace functions by multifunctions. A natural conjecture, in light of the above, is the following: Conjecture. Let \( S \) in \( {\mathbb{R}}^{n} \) be homeomorphic to a closed unit ball, and let \( F \) be a locally Lipschitz multifunction from \( {\mathbb{R}}^{n} \) to \( {\mathbb{R}}^{n} \) such that \( \left( {S, F}\right) \) is weakly invariant. Then \( F \) has a zero in \( S \) . ![f81c7d92-0efa-420d-b97e-e67e14ba6fc3_215_0.jpg](images/f81c7d92-0efa-420d-b97e-e67e14ba6fc3_215_0.jpg) FIGURE 4.2. The set \( S \) of Exercise 4.3. Natural, but false! Let us construct a counterexample. \( S \) is the following subset of \( {\mathbb{R}}^{3} \) (a "fluted funnel"; see Figure 4.2): \[ \left\{ {\left( {x, y, z}\right) : {x}^{2} + {y}^{2} = {z}^{4},0 \leq z \leq 1}\right\} . \] We define \( f : {\mathbb{R}}^{3} \rightarrow {\mathbb{R}}^{3} \) as follows: \[ f\left( {x, y, z}\right) \mathrel{\text{:=}} \left\lbrack {-{yz} + \frac{{2x}\left( {1 - z}\right) }{{\left( {x}^{2} + {y}^{2}\right) }^{1/4}},{xz} + \frac{{2y}\left( {1 - z}\right) }{{\left( {x}^{2} + {y}^{2}\right) }^{1/4}},1 - z}\right\rbrack \] for \( {x}^{2} + {y}^{2} \neq 0 \), and \[ f\left( {0,0, z}\right) \mathrel{\text{:=}} \left\lbrack {0,0,1 - z}\right\rbrack . \] 4.3. Exercise. (a) \( S \) is homeomorphic to the closed unit ball in \( {\mathbb{R}}^{2} \), and \( S \) is regular. (b) \( f \) is continuous everywhere, not Lipschitz at the origin, and \( f \) has no zero in \( S \) . (c) For all \( \left( {x, y, z}\right) \in S \), we have \( f\left( {x, y, z}\right) \in {T}_{S}^{B}\left( {x, y, z}\right) \) . Deduce that the system \( \left( {S, f}\right) \) is weakly invariant, but show that it is not strongly invariant. (Figure 4.2 indicates some trajectories of \( f \), including two that start at 0, one which remains in the set \( S \), while the other leaves it.) The preceding exercise shows that our conjecture is false when the multifunction \( F \) is a continuous (single-valued) function. However, the exercise does not directly address the case of a Lipschitz multifunction. That gap will be bridged by a useful result which asserts that upper semicontinuous multifunctions can be approximated "from above" by Lipschitz ones. 4.4. Proposition. Let \( F \) satisfy the Standing Hypotheses. Then there exists a sequence of locally Lipschitz multifunctions \( \left\{ {F}_{k}\right\} \) also satisfying the Standing Hypotheses such that: (i) For each \( k \in \mathbb{N} \), for every \( x \in {\mathbb{R}}^{n} \) , \[ F\left( x\right) \subseteq {F}_{k + 1}\left( x\right) \subseteq {F}_{k}\left( x\right) \subseteq \overline{\operatorname{co}}F\left( {B\left( {x;{3}^{-k + 1}}\right) }\right) . \] (ii) \( \mathop{\bigcap }\limits_{{k \geq 1}}{F}_{k}\left( x\right) = F\left( x\right) \forall x \in {\mathbb{R}}^{n} \) . Proof. Let us fix an integer \( k \geq 1 \) and consider the set of all points (or grid) in \( {\mathbb{R}}^{n} \) whose coordinates are all of the form \( \pm m/{4}^{k}, m \in \{ 0,1,2,\ldots \} \) . Let \( \left\{ {x}_{i}\right\} \) be an enumeration of this countable set \( \left( {i = 1,2,\ldots }\right) \), and observe that the collection of open balls \( B\left( {{x}_{i};{3}^{-k}}\right) \) of radius \( {3}^{-k} \) around \( {x}_{i}\left( {i = 1,2,\ldots }\right) \) is a covering of \( {\mathbb{R}}^{n} \) . This covering has the property that any bounded set intersects only finitely many members of the collection. This implies the existence of a locally Lipschitz partition of unity for the covering; that is, a collection of functions \( {p}_{i}\left( x\right) \), each locally Lipschitz and taking values in \( \left\lbrack {0,1}\right\rbrack \), with the following properties: \[ x \notin B\left( {{x}_{i};{3}^{-k}}\right) \Rightarrow {p}_{i}\left( x\right) = 0\forall i, \] \[ \mathop{\sum }\limits_{i}{p}_{i}\left( x\right) = 1\forall x \] Note that only finitely many terms in the last sum are nonzero, so there is no difficulty regarding its interpretation. 4.5. Exercise. Show that the desired Lipschitz partition of unity is obtained by setting \[ {p}_{i}\left( x\right) \mathrel{\text{:=}} \frac{d\left( {x,\operatorname{comp}{B}_{i}^{k}}\right) }{\mathop{\sum }\limits_{j}d\left( {x,\operatorname{comp}{B}_{j}^{k}}\right) }, \] where \( {B}_{i}^{k} \mathrel{\text{:=}} B\left( {{x}_{i};{3}^{-k}}\right) \) . Returning to the proof, we define the compact convex set \[ {C}_{i}^{k} \mathrel{\text{:=}} \overline{\operatorname{co}}F\left( {B\left( {{x}_{i};2 \cdot {3}^{-k}}\right) }\right) \] and the multifunction \( {F}_{k} \) via \[ {F}_{k}\left( x\right) \mathrel{\text{:=}} \mathop{\sum }\limits_{i}{p}_{i}^{k}\left( x\right) {C}_{i}^{k} \] where the dependence of \( {p}_{i} \) on \( k \) has been reflected by the notation \( {p}_{i}^{k} \) . Then \( {F}_{k} \) is locally Lipschitz, as is easily seen. In addition, \( {F}_{k}\left( x\right) \) contains \( F\left( x\right) \) . To see this, note that for given \( x \), those terms actually contributing in the sum defining \( {F}_{k}\left( x\right) \) have \( {p}_{i}^{k}\left( x\right) > 0 \), whence \( x \in {B}_{i}^{k} \) . But for such "active indices" \( i \), we then have \( F\left( x\right) \subset {C}_{i}^{k} \) . Since the coefficients \( {p}_{i}^{k}\left( x\right) \) taken over the active indices form a convex combination, it follows that \( {F}_{k}\left( x\right) \) contains \( F\left( x\right) \) . Also, when \( x \in {B}_{i}^{k} \) we have reciprocally that \( {x}_{i} \in B\left( {x;{3}^{-k}}\right) \), whence \( B\left( {{x}_{i},2 \cdot {3}^{-k}}\right) \subset B\left( {x,{3}^{-k + 1}}\right) \) . Thus, for each such \( i,{C}_{i}^{k} \) is contained in \( \overline{\operatorname{co}}F\left( {B\left( {x;{3}^{-k + 1}}\right) }\right) \), and so is \( {F}_{k}\left( x\right) \) . This confirms the last estimate in (i), which clearly implies (ii) in light of the upper semicontinuity of \( F \) . There remains only the decreasing nature of the approximation to check. Letting \( {F}_{k + 1}\left( x\right) = \mathop{\sum }\limits_{i}{p}_{i}^{k + 1}\left( x\right) {C}_{i}^{k + 1} \), consider any two indices \( i \) and \( j \) which are active for \( {F}_{k + 1}\left( x\right) \) and \( {F}_{k}\left( x\right) \), respectively (these are the only ones that actually matter). Then \( {p}_{i}^{k + 1}\left( x\right) > 0,{p}_{j}^{k}\left( x\right) > 0 \), whence \( x \in \) \( B\left( {{x}_{i},{3}^{-k - 1}}\right) \cap B\left( {{x}_{j},{3}^{-k}}\right) \), which implies \( \begin{Vmatrix}{{x}_{i} - {x}_{j}}\end{Vmatrix} < {3}^{-k - 1} + {3}^{-k} \), which in turn implies \( {C}_{i}^{k + 1} \subset {C}_{j}^{k} \) . This shows that each set contributing to the sum defining \( {F}_{k + 1}\left( x\right) \) is contained in every set in the sum defining \( {F}_{k}\left( x\right) \) , which completes the proof. Now let us return to the burial of the conjecture. Applying Proposition 4.4 to the multifunction \( F\left( x\right) \mathrel{\text{:=}} \{ f\left( x\right) \} \), where \( f \) is the function of Exercise 4.3,
1012_(GTM178)Nonsmooth Analysis and Control Theory
73
its_{i}{p}_{i}^{k + 1}\left( x\right) {C}_{i}^{k + 1} \), consider any two indices \( i \) and \( j \) which are active for \( {F}_{k + 1}\left( x\right) \) and \( {F}_{k}\left( x\right) \), respectively (these are the only ones that actually matter). Then \( {p}_{i}^{k + 1}\left( x\right) > 0,{p}_{j}^{k}\left( x\right) > 0 \), whence \( x \in \) \( B\left( {{x}_{i},{3}^{-k - 1}}\right) \cap B\left( {{x}_{j},{3}^{-k}}\right) \), which implies \( \begin{Vmatrix}{{x}_{i} - {x}_{j}}\end{Vmatrix} < {3}^{-k - 1} + {3}^{-k} \), which in turn implies \( {C}_{i}^{k + 1} \subset {C}_{j}^{k} \) . This shows that each set contributing to the sum defining \( {F}_{k + 1}\left( x\right) \) is contained in every set in the sum defining \( {F}_{k}\left( x\right) \) , which completes the proof. Now let us return to the burial of the conjecture. Applying Proposition 4.4 to the multifunction \( F\left( x\right) \mathrel{\text{:=}} \{ f\left( x\right) \} \), where \( f \) is the function of Exercise 4.3, gives rise to a sequence of locally Lipschitz multifunctions \( {F}_{k} \) such that \( \left( {S,{F}_{k}}\right) \) is weakly invariant for each \( k \) ; yet \( {F}_{k} \) cannot admit a zero in \( S \) for \( k \) arbitrarily large, otherwise \( f \) would do so as well. To summarize, the condition \[ F\left( x\right) \cap {T}_{S}^{B}\left( x\right) \neq \varnothing \;\left( {x \in S}\right) \] although it characterizes weak invariance, is not adequate (even when \( F \) is Lipschitz) to yield a zero of \( F \), even when \( S \) is homeomorphic to the closed unit ball (though it is adequate when \( S \) is a closed unit ball, as will follow from results below). ## Existence of Equilibria in the Multivalued Case Having seen what fails to be true, let us proceed to derive a positive result; the key is to use \( {T}_{S}^{C} \) instead of \( {T}_{S}^{B} \) . 4.6. Theorem. Let \( S \) be homeomorphic to a closed unit ball in \( {\mathbb{R}}^{n} \), and let \( F \) satisfy the Standing Hypotheses. Suppose that: (i) \( \operatorname{int}{T}_{S}^{C}\left( x\right) \neq \varnothing \forall x \in S \) (i.e., \( S \) is wedged). (ii) \( F\left( x\right) \cap {T}_{S}^{C}\left( x\right) \neq \varnothing \forall x \in S \) . Then \( F \) has a zero in \( S \) . Even if \( F \) is Lipschitz, this fails in general in the absence of (i) (even when \( S \) is regular), or if \( {T}_{S}^{B}\left( x\right) \) replaces \( {T}_{S}^{C}\left( x\right) \) in (ii). Proof. We make the temporary hypothesis that \( F \) is locally Lipschitz and satisfies \[ F\left( x\right) \cap \operatorname{int}{T}_{S}^{C}\left( x\right) \neq \varnothing \forall x \in S. \] Let \( x \in S \), and let \( \varepsilon > 0 \) be given. Then, because \( F \) is Lipschitz near \( x \) , and because \( {T}_{S}^{C} \) is lower semicontinuous at \( x \) (by Proposition 3.6.7), there exists for each \( y \in F\left( x\right) \cap \operatorname{int}{T}_{S}^{C}\left( x\right) \) a scalar \( \delta \left( {x, y}\right) > 0 \) such that \[ {x}^{\prime } \in x + \delta \left( {x, y}\right) B \Rightarrow y \in \left\{ {F\left( {x}^{\prime }\right) + {\varepsilon B}}\right\} \cap \operatorname{int}{T}_{S}^{C}\left( {x}^{\prime }\right) . \] The family of open balls \( x + \delta \left( {x, y}\right) B \) forms an open covering of the compact set \( S \) ; let the family of sets \( \left\{ {{x}_{i} + \delta \left( {{x}_{i},{y}_{i}}\right) B}\right\} \left( {i = 1,2,\ldots, m}\right) \) be a finite subcover. Now associate to this subcover a Lipschitz partition of unity \( \left\{ {{p}_{i}\left( x\right) }\right\} \) subordinate to it, and define a locally Lipschitz function \( {f}_{\varepsilon } \) as follows: \[ {f}_{\varepsilon }\left( x\right) \mathrel{\text{:=}} \mathop{\sum }\limits_{{i = 1}}^{m}{p}_{i}\left( x\right) {y}_{i} \] When \( {p}_{i}\left( x\right) \neq 0 \), we have \( x \in {x}_{i} + \delta \left( {{x}_{i},{y}_{i}}\right) B \), whence \[ {y}_{i} \in \{ F\left( x\right) + {\varepsilon B}\} \cap \operatorname{int}{T}_{S}^{C}\left( x\right) \] by definition of \( \delta \left( {{x}_{i},{y}_{i}}\right) \) . Because \( F\left( x\right) + {\varepsilon B} \) and int \( {T}_{S}^{C}\left( x\right) \) are both convex sets, we infer \[ {f}_{\varepsilon }\left( x\right) \in \{ F\left( x\right) + {\varepsilon B}\} \cap \operatorname{int}{T}_{S}^{C}\left( x\right) . \] Thus \( {f}_{\varepsilon }\left( x\right) \in {T}_{S}^{B}\left( x\right) \), so that the system \( \left( {S,{f}_{\varepsilon }}\right) \) is weakly invariant by Theorem 2.10. By Exercise 4.2, there is a point \( {x}_{\varepsilon } \in S \) such that \( {f}_{\varepsilon }\left( {x}_{\varepsilon }\right) = 0 \) , which implies \[ 0 \in F\left( {x}_{\varepsilon }\right) + {\varepsilon B}\text{.} \] Now let \( \varepsilon \downarrow 0 \) . Some subsequence \( \left\{ {x}_{{\varepsilon }_{i}}\right\} \) of the corresponding points converges to a limit \( {x}^{ * } \in S \), and the upper semicontinuity of \( F \) yields \( 0 \in \) \( F\left( {x}^{ * }\right) \) . Thus the theorem is proven, under the temporary hypothesis. To remove the need for this hypothesis, let us observe first that it is satisfied by each of the multifunctions \[ {\widetilde{F}}_{k}\left( x\right) \mathrel{\text{:=}} {F}_{k}\left( x\right) + \gamma \bar{B} \] together with the hypotheses of the theorem under consideration, where \( {F}_{k} \) is the Lipschitz approximation to \( F \) provided by Proposition 4.4, and where \( \gamma > 0 \) is fixed. By the case of the theorem proved above, there exists \( {x}_{k} \in S \) such that \( 0 \in {\widetilde{F}}_{k}\left( {x}_{k}\right) \), and by passing to a subsequence (without relabeling) we may assume \( {x}_{k} \rightarrow {x}^{ * } \in S \) . Now the monotonicity of the \( {F}_{k} \) scheme gives us: \[ 0 \in {F}_{k}\left( {x}_{j}\right) + \gamma \bar{B}\forall j \geq k. \] Letting \( j \rightarrow \infty \) yields \( 0 \in {F}_{k}\left( {x}^{ * }\right) + \gamma \bar{B} \), since \( {F}_{k} \) is Lipschitz and all the more upper semicontinuous. We invoke (ii) of Proposition 4.4 to deduce \( 0 \in \) \( F\left( {x}^{ * }\right) + \gamma \bar{B} \) . Since \( \gamma \) is an arbitrary positive number, an evident sequential argument together with the upper semicontinuity of \( F \) confirms that \( F \) has a zero in \( S \) . There remain to prove the two negative assertions in the statement of the theorem. The first of these is confirmed by the example above that buried the conjecture, while the second follows from the example in Exercise 2.12 (see Exercise 4.8 below). Under the right hypotheses, nonconvex sets which are weakly invariant necessarily contain equilibria, which is the assertion of the next corollary. 4.7. Corollary. Let \( S \) be homeomorphic to the closed unit ball in \( {\mathbb{R}}^{n} \) . Further assume that \( S \) is wedged and regular, and that the system \( \left( {S, F}\right) \) is weakly invariant, where \( F \) satisfies the Standing Hypotheses. Then \( S \) contains a zero of \( F \) . 4.8. Exercise. (a) Show that for a suitable locally Lipschitz approximation \( {F}_{k} \) of the multifunction \( F \) appearing in Exercise 2.12, \( {F}_{k} \) together with the set \( S \) of that exercise satisfies all the hypotheses of Theorem 4.6, but with \( {T}_{S}^{C} \) replaced by \( {T}_{S}^{B} \) . Yet \( {F}_{k} \) has no zero in \( S \) . (Hint. \( S \) is star-shaped.) (b) Prove Corollary 4.7. ## 5 Lyapounov Theory and Stabilization Consider the differential equation \[ \dot{x}\left( t\right) = f\left( {x\left( t\right) }\right) \] where \( f \) is a smooth function from \( {\mathbb{R}}^{n} \) to itself, and suppose that the point \( {x}^{ * } \) is an equilibrium: \( f\left( {x}^{ * }\right) = 0 \) . Then of course the constant function \( x\left( t\right) \equiv {x}^{ * } \) is a solution of the differential equation. If for any \( \alpha \in {\mathbb{R}}^{n} \) the solution \( x\left( \cdot \right) \) of the differential equation satisfying \( x\left( 0\right) = \alpha \) exists on \( \lbrack 0,\infty ) \) and has the property that \( x\left( t\right) \rightarrow {x}^{ * } \) as \( t \rightarrow \infty \), then the equilibrium is said to be globally asymptotically stable. We recall that if \( \alpha \neq {x}^{ * } \), then \( {x}^{ * } \) cannot be reached in finite time: \( x\left( t\right) \neq {x}^{ * }\forall t > 0 \) . A simple but far-reaching criterion to assure this asymptotic stability can be given in terms of Lyapounov functions. Suppose that smooth functions \( Q \) and \( W \) exist such that: (i) \( Q\left( x\right) \geq 0, W\left( x\right) \geq 0\forall x \in {\mathbb{R}}^{n} \), and \( W\left( x\right) = 0 \) iff \( x = {x}^{ * } \) ; (ii) the sets \( \{ x : Q\left( x\right) \leq q\} \) are compact \( \forall q \in \mathbb{R} \) ; and (iii) \( \langle \nabla Q\left( x\right), f\left( x\right) \rangle \leq - W\left( x\right) \forall x \) . (These properties are referred to as positive definiteness, growth, and infinitesimal decrease, respectively.) It follows that \( {x}^{ * } \) must be globally asymptotically stable. For let \( x\left( \cdot \right) \) be any local solution of the differential equation. We have \[ \frac{d}{dt}Q\left( {x\left( t\right) }\right) = \langle \nabla Q\left( {x\left( t\right) }\right), f\left( {x\left( t\right) }\right) \rangle \leq - W\left( {x\left( t\right) }\right) \] by condition (iii). It follows that the nonnegative function \[ t \mapsto Q\left( {x\left( t\right) }\right) + {\int }_{0}^{t}W\left( {x\left( \tau \right) }\right) {d\tau } \] is decreasing and hence bounded on \( \lbrack 0,\infty ) \) . Since \( W \geq 0 \), this implies that the function \( t \mapsto Q\left( {x\left( t\right) }\right) \) is bounded and hence that \( x\left( t\right) \) remains bounded. But then the solution \( x\left( t\right) \) of the differential equation exists on \( \lbrack 0,\infty ) \),
1012_(GTM178)Nonsmooth Analysis and Control Theory
74
x \) . (These properties are referred to as positive definiteness, growth, and infinitesimal decrease, respectively.) It follows that \( {x}^{ * } \) must be globally asymptotically stable. For let \( x\left( \cdot \right) \) be any local solution of the differential equation. We have \[ \frac{d}{dt}Q\left( {x\left( t\right) }\right) = \langle \nabla Q\left( {x\left( t\right) }\right), f\left( {x\left( t\right) }\right) \rangle \leq - W\left( {x\left( t\right) }\right) \] by condition (iii). It follows that the nonnegative function \[ t \mapsto Q\left( {x\left( t\right) }\right) + {\int }_{0}^{t}W\left( {x\left( \tau \right) }\right) {d\tau } \] is decreasing and hence bounded on \( \lbrack 0,\infty ) \) . Since \( W \geq 0 \), this implies that the function \( t \mapsto Q\left( {x\left( t\right) }\right) \) is bounded and hence that \( x\left( t\right) \) remains bounded. But then the solution \( x\left( t\right) \) of the differential equation exists on \( \lbrack 0,\infty ) \), and \( W\left( {x\left( t\right) }\right) \) is a globally Lipschitz function on \( \lbrack 0,\infty ) \) . That \( W\left( {x\left( t\right) }\right) \) converges to 0 now follows from: 5.1. Exercise. Let \( r\left( t\right) \) be a nonnegative, globally Lipschitz function on \( \lbrack 0,\infty ) \), and suppose that \( {\int }_{0}^{t}r\left( \tau \right) {d\tau } \) is bounded for \( t \geq 0 \) . Then \( r\left( t\right) \rightarrow 0 \) as \( t \rightarrow \infty \) . Knowing that \( W\left( {x\left( t\right) }\right) \rightarrow 0 \), it follows readily that \( x\left( t\right) \) converges to \( {x}^{ * } = \) \( {W}^{-1}\left( 0\right) \) as \( t \rightarrow \infty \), as claimed. Note the prominence of monotonicity in this classical argument; it will play a central role in the results of this section as well. The method under discussion here was introduced by A. M. Lyapounov in the theory of differential equations. Its extension to control systems is conditioned by the fact that we typically consider situations in which some but not all trajectories from a given initial condition approach the equilibrium. That is, an element of "controllability" is involved. With this in mind, here is a natural way to proceed in the case of a differential inclusion \( \dot{x} \in F\left( x\right) \) admitting an equilibrium at \( {x}^{ * } \) (i.e., such that \( 0 \in F\left( {x}^{ * }\right) \) ). We define a (smooth) Lyapounov pair \( \left( {Q, W}\right) \) to be one having the same properties (i) and (ii) as above, together with the following replacement for (iii): \( {\left( iii\right) }^{\prime }\mathop{\min }\limits_{{v \in F\left( x\right) }}\langle \nabla Q\left( x\right), v\rangle \leq - W\left( x\right) \forall x. \) This does the trick, as we ask the reader to check. 5.2. Exercise. If \( Q \) and \( W \) are \( {C}^{1} \) and satisfy (i),(ii), and (iii)', then for any \( \alpha \in {\mathbb{R}}^{n} \) there is a trajectory \( x \) on \( \lbrack 0,\infty ) \) with \( x\left( 0\right) = \alpha \) such that \( x\left( t\right) \rightarrow {x}^{ * } \) as \( t \rightarrow \infty \) . (Hint. Consider trajectories of the multifunction \[ \widetilde{F}\left( x\right) \mathrel{\text{:=}} \{ v \in F\left( x\right) : \langle \nabla Q\left( x\right), v\rangle \leq - W\left( x\right) \} .) \] We remark that the very existence of \( \left( {Q, W}\right) \) implies that \( {x}^{ * } \) is an equilibrium, and that in contrast to the classical case of a differential equation, \( x\left( \cdot \right) \) may reach \( {x}^{ * } \) in finite time. One of the interesting questions about this approach is whether it is necessary as well as sufficient for asymptotic stability. In this connection, and for other reasons as well, it turns out to be essential to develop a theory encompassing nonsmooth Lyapounov functions. In light of the earlier chapters, our proposal in this direction is the following: take \( Q \) and \( W \) in the class \( \mathcal{F}\left( {\mathbb{R}}^{n}\right) \), retain properties (i) and (ii), and add to them the following: for every \( x \), for every \( \zeta \in {\partial }_{P}Q\left( x\right) ,\exists v \in F\left( x\right) \) such that \( \langle \zeta, v\rangle \leq - W\left( x\right) \) . We summarize, and in so doing express this proximal form of the infinitesimal decrease property in terms of the lower Hamiltonian \( h \) . Two functions \( Q \) and \( W \) in \( \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) are said to be a Lyapounov pair for \( {x}^{ * } \) if they satisfy the following properties: - Positive definiteness: \( Q, W \geq 0;W\left( x\right) = 0 \) iff \( x = {x}^{ * } \) . - Growth: the level sets \( \left\{ {x \in {\mathbb{R}}^{n} : Q\left( x\right) \leq q}\right\} \) are compact \( \forall q \in \mathbb{R} \) . - Infinitesimal decrease: \( h\left( {x,{\partial }_{P}Q\left( x\right) }\right) \leq - W\left( x\right) \forall x \in {\mathbb{R}}^{n} \) . Note that this last property holds automatically at \( x = {x}^{ * } \) when \( {x}^{ * } \) is an equilibrium, since then \( 0 \in F\left( {x}^{ * }\right) \) . Furthermore, infinitesimal decrease can also be expressed in an equivalent form using the subderivates of \( §{3.4} \) : 5.3. Proposition. We have \( h\left( {x,{\partial }_{P}Q\left( x\right) }\right) \leq - W\left( x\right) \forall x \in {\mathbb{R}}^{n} \) iff \[ \mathop{\inf }\limits_{{v \in F\left( x\right) }}{DQ}\left( {x;v}\right) \leq - W\left( x\right) \forall x \in \operatorname{dom}Q. \] Proof. Suppose first that the derivate condition holds. Let \( \zeta \in {\partial }_{P}Q\left( x\right) \) , and pick \( v \in F\left( x\right) \) such that \( {DQ}\left( {x;v}\right) \leq - W\left( x\right) \) (we know from Exercise 3.4.1 that \( {DQ}\left( {x; \cdot }\right) \) is lower semicontinuous, so the infimum is attained). Since \( {DQ}\left( {x;v}\right) \geq \langle \zeta, v\rangle \) in general, we have \( \langle \zeta, v\rangle \leq - W\left( x\right) \) , whence \( h\left( {x,{\partial }_{P}Q\left( x\right) }\right) \leq - W\left( x\right) \) ; i.e., the Hamiltonian inequality holds. The necessity is a deeper result. If the derivate condition fails, then for some \( \delta > 0 \) we have \[ {DQ}\left( {x;v}\right) \geq - W\left( x\right) + \delta \forall v \in F\left( x\right) . \] Since \( {DQ}\left( {x; \cdot }\right) \) is lower semicontinuous and \( F\left( x\right) \) is compact, this implies that for some \( \eta > 0 \) we have \[ {DQ}\left( {x;v}\right) > - W\left( x\right) + \frac{\delta }{2}\forall v \in F\left( x\right) + \eta \bar{B}. \] Applying Subbotin’s Theorem 3.4.2, we deduce that for some \( z \) arbitrarily near \( x \), for some \( \zeta \in {\partial }_{P}Q\left( z\right) \), we have \[ \langle \zeta, v\rangle > - W\left( x\right) + \frac{\delta }{2}\forall v \in F\left( x\right) + \eta \bar{B} \] For \( z \) near enough to \( x \), we have \( F\left( z\right) \subset F\left( x\right) + {\eta B} \), since \( F \) is upper semicontinuous, as well as \( W\left( z\right) \geq W\left( x\right) - \delta /4 \), since \( W \) is lower semicontinuous. For such \( z \) we deduce \[ \langle \zeta, v\rangle > - W\left( z\right) + \frac{\delta }{4}\forall v \in F\left( z\right) . \] This implies \( h\left( {z,\zeta }\right) \geq - W\left( z\right) + \delta /4 \), which shows that the Hamiltonian condition fails at \( z \), thereby completing the proof. It turns out that without any loss of generality, we can always assume that \( W \) has additional properties, in particular, being finite-valued. However, the fact that \( Q \) can be extended-valued is useful, for example, in subsuming local stability, and is to be retained. ## 5.4. Exercise. (a) If \( Q \) and \( W \) in \( \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) form a Lyapounov pair for \( {x}^{ * } \), prove that there is a function \( \widetilde{W} \) for which \( \left( {Q,\widetilde{W}}\right) \) continues to be a Lyapounov pair for \( {x}^{ * } \), where \( \widetilde{W} \) is locally Lipschitz and satisfies a linear growth condition. (Hint. \[ \widetilde{W}\left( x\right) \mathrel{\text{:=}} \min \left\{ {W\left( y\right) + \parallel x - y\parallel : y \in {\mathbb{R}}^{n}}\right\} .) \] (b) Prove that \( Q\left( x\right) > 0 \) for any \( x \neq {x}^{ * } \) . (Hint. Otherwise, \( 0 \in \) \( \left. {{\partial }_{P}Q\left( x\right) \text{.}}\right) \) (c) Given \( x \neq {x}^{ * } \), there exist points \( y \) arbitrarily near \( {x}^{ * } \) having \( Q\left( y\right) < Q\left( x\right) \) . (d) \( Q \) attains a global minimum at \( {x}^{ * } \) (which we can always take to be 0 by redefining \( Q \) if necessary). 5.5. Theorem. Let \( 0 \in F\left( {x}^{ * }\right) \), and let there exist \( Q \) and \( W \) in \( \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) such that \( \left( {Q, W}\right) \) constitutes a Lyapounov pair for \( {x}^{ * } \) . Then for any \( \alpha \in \operatorname{dom}Q \) there is a trajectory \( x \) for \( F \) on \( \lbrack 0,\infty ) \) having \( x\left( 0\right) = \alpha \) such that \( x\left( t\right) \rightarrow {x}^{ * } \) as \( t \rightarrow \infty \) . The proof of this important result requires a proximal characterization of a certain system monotonicity property, and is postponed until after that acquisition. ## Weakly Decreasing Systems Let \( \varphi \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) . The system \( \left( {\varphi, F}\right) \) is said to be weakly decreasing if for any \( \alpha \in {\mathbb{R}}^{n} \), there exists a trajectory \( x \) of \( F \) on \( \lbrack 0,\infty ) \) with \( x\left( 0\right) = \alpha \) which satisfies \[ \varphi \left( {x\left( t\right) }\right) \leq \varphi \left( {x\left( 0\right) }\right) = \varphi \left( \alpha \right) \forall t \geq 0. \] Note that only points \( \alpha \in \operatorname{dom}\varphi \) need be tested in this condition, since it holds automatically otherwise. We
1012_(GTM178)Nonsmooth Analysis and Control Theory
75
Q \) there is a trajectory \( x \) for \( F \) on \( \lbrack 0,\infty ) \) having \( x\left( 0\right) = \alpha \) such that \( x\left( t\right) \rightarrow {x}^{ * } \) as \( t \rightarrow \infty \) . The proof of this important result requires a proximal characterization of a certain system monotonicity property, and is postponed until after that acquisition. ## Weakly Decreasing Systems Let \( \varphi \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) . The system \( \left( {\varphi, F}\right) \) is said to be weakly decreasing if for any \( \alpha \in {\mathbb{R}}^{n} \), there exists a trajectory \( x \) of \( F \) on \( \lbrack 0,\infty ) \) with \( x\left( 0\right) = \alpha \) which satisfies \[ \varphi \left( {x\left( t\right) }\right) \leq \varphi \left( {x\left( 0\right) }\right) = \varphi \left( \alpha \right) \forall t \geq 0. \] Note that only points \( \alpha \in \operatorname{dom}\varphi \) need be tested in this condition, since it holds automatically otherwise. We will see that this system property is the functional counterpart of weak invariance of sets. 5.6. Exercise. (a) Let \( \varphi \) be the indicator of a closed set \( S \) . Show then that \( \left( {\varphi, F}\right) \) is weakly decreasing iff \( \left( {S, F}\right) \) is weakly invariant. (b) Show that \( \left( {\varphi, F}\right) \) is weakly decreasing iff \( \left( {\operatorname{epi}\varphi, F\times \{ 0\} }\right) \) is weakly invariant. ## 5.7. Theorem. \( \left( {\varphi, F}\right) \) is weakly decreasing iff \[ h\left( {x,{\partial }_{P}\varphi \left( x\right) }\right) \leq 0\forall x \in {\mathbb{R}}^{n}. \] Proof. Note the meaning of this Hamiltonian inequality: for any \( x \in {\mathbb{R}}^{n} \) , and \( \zeta \in {\partial }_{P}\varphi \left( x\right) \), we have \( h\left( {x,\zeta }\right) \leq 0 \) . Suppose first that \( \left( {\varphi, F}\right) \) is weakly decreasing. Then by Exercise 5.6, \( \left( {\operatorname{epi}\varphi, F\times \{ 0\} }\right) \) is weakly invariant. In light of Theorem 2.4, this is characterized by the condition that for any vector \( \left( {\zeta ,\lambda }\right) \in {N}_{\text{epi }\varphi }^{P}\left( {x, r}\right) \), where \( \left( {x, r}\right) \in \operatorname{epi}\varphi \), we have \[ \min \{ \left( {\zeta ,\lambda }\right) \cdot \left( {v,0}\right) : v \in F\left( x\right) \} \leq 0. \] Now if \( \zeta \in {\partial }_{P}\varphi \left( x\right) \), then \( \left( {\zeta , - 1}\right) \in {N}_{\text{epi }\varphi }^{P}\left( {x,\varphi \left( x\right) }\right) \), and so we deduce \[ h\left( {x,\zeta }\right) = \min \{ \langle \zeta \cdot v\rangle : v \in F\left( x\right) \} \leq 0, \] which confirms the Hamiltonian inequality. For the converse, suppose that the Hamiltonian condition holds. In order to deduce that \( \left( {\varphi, F}\right) \) is weakly decreasing, or equivalently that \( \left( {\operatorname{epi}\varphi, F\times \{ 0\} }\right) \) is weakly invariant, it suffices to exhibit for any \( \left( {\zeta ,\lambda }\right) \in {N}_{\text{epi }\varphi }^{P}\left( {x, r}\right) \) an element \( v \in F\left( x\right) \) such that \( \langle \zeta, v) \leq 0 \) . We know that \( \lambda \leq {0}^{\prime } \) and that \( \left( {\zeta ,\lambda }\right) \in {N}_{\text{epi }\varphi }^{P}\left( {x,\varphi \left( x\right) }\right) \), by Exercise 1.2.1(d). If \( \lambda < 0 \), then we have \[ \left( {\zeta /\left( {-\lambda }\right) , - 1}\right) \in {N}_{\text{epi }\varphi }^{P}\left( {x, r}\right) \] which implies \( - \zeta /\lambda \in {\partial }_{P}\varphi \left( x\right) \) . Then the Hamiltonian condition applies to imply the existence of \( v \in F\left( x\right) \) for which \( \langle \left( {-\zeta /\lambda }\right), v\rangle \leq 0 \) . But then \( \langle \zeta, v\rangle \leq 0 \) as required. The remaining case to consider is that in which \( \lambda = 0 \) . Then we have \[ \left( {\zeta ,0}\right) \in {N}_{\text{epi }\varphi }^{P}\left( {x,\varphi \left( x\right) }\right) \] and we invoke Problem 1.11.23 to deduce the existence of sequences \( \left( {{\zeta }_{i}, - {\varepsilon }_{i}}\right) \) , with \( {\varepsilon }_{i} > 0 \), and \( \left( {{x}_{i},\varphi \left( {x}_{i}\right) }\right) \) such that \[ \left( {{\zeta }_{i}, - {\varepsilon }_{i}}\right) \rightarrow \left( {\zeta ,0}\right) ,\;\left( {{\zeta }_{i}, - {\varepsilon }_{i}}\right) \in {N}_{\operatorname{epi}\varphi }^{P}\left( {{x}_{i},\varphi \left( {x}_{i}\right) }\right) ,\;\left( {{x}_{i},\varphi \left( {x}_{i}\right) }\right) \rightarrow \left( {x,\varphi \left( x\right) }\right) . \] Then, as in the case \( \lambda < 0 \) above, there exists \( {v}_{i} \in F\left( {x}_{i}\right) \) for which \( \left\langle {{\zeta }_{i},{v}_{i}}\right\rangle \leq \) 0 . Since \( F \) is locally bounded, the sequence \( \left\{ {v}_{i}\right\} \) is bounded. Passing to a subsequence, we can suppose that \( {v}_{i} \) converges to a limit \( v \) ; then \( v \in F\left( x\right) \) as a consequence of the upper semicontinuity of \( F \) . We deduce \( \langle \zeta, v\rangle \leq 0 \) , which completes the proof. ## Proof of Theorem 5.5 Let \( \left( {Q, W}\right) \) be a Lyapounov pair for \( {x}^{ * } \), where \( W \) has the additional properties provided by Exercise 5.4, and let \( \alpha \in \operatorname{dom}Q \) . We will prove the existence of the trajectory \( x \) cited in the statement of the theorem. We define \( \widetilde{Q} : {\mathbb{R}}^{n} \times \mathbb{R} \rightarrow ( - \infty ,\infty \rbrack \) via \( \widetilde{Q}\left( {x, y}\right) \mathrel{\text{:=}} Q\left( x\right) + y \), and the multifunction \( \widetilde{F} \) as follows: \[ \widetilde{F}\left( {x, y}\right) \mathrel{\text{:=}} F\left( x\right) \times \{ W\left( x\right) \} . \] Note that \( \widetilde{F} \) satisfies the Standing Hypotheses. We claim that the system \( \left( {\widetilde{Q},\widetilde{F}}\right) \) is weakly decreasing. Indeed, with an eye to applying Theorem 5.7, let \( \left( {\zeta, r}\right) \in {\partial }_{P}\widetilde{Q}\left( {x, y}\right) \) . Then \( \zeta \in {\partial }_{P}Q\left( x\right) \) and \( r = 1 \) . The infinitesimal decrease property of \( \left( {Q, W}\right) \) provides the existence of \( v \in F\left( x\right) \) such that \( \langle \left( {v, W\left( x\right) }\right) ,\left( {\zeta ,1}\right) \rangle \leq 0 \), which verifies the Hamiltonian inequality of Theorem 5.7. We deduce the existence of a trajectory \( \left( {x, y}\right) \) of \( \widetilde{F} \) beginning at \( \left( {\alpha ,0}\right) \) and satisfying \( \widetilde{Q}\left( {x\left( t\right), y\left( t\right) }\right) \leq \widetilde{Q}\left( {\alpha ,0}\right) = Q\left( \alpha \right) \) for \( t \geq 0 \) . This translates to \[ Q\left( {x\left( t\right) }\right) + {\int }_{0}^{t}W\left( {x\left( \tau \right) }\right) {d\tau } \leq Q\left( \alpha \right) \] where \( x \) is a trajectory of \( F \) . This implies that \( Q\left( {x\left( t\right) }\right) \) (and hence \( x\left( t\right) \) ) is bounded, as well as \( {\int }_{0}^{t}W\left( {x\left( \tau \right) }\right) {d\tau } \) . Since \( F \) is bounded on bounded sets, we observe that \( \dot{x}\left( t\right) \) is bounded too, whence \( x \) satisfies a global Lipschitz condition on \( \lbrack 0,\infty ) \), let us say of rank \( k \) . We conclude by proving what amounts to Exercise 5.1. Suppose that \( x\left( t\right) \) fails to converge to \( {x}^{ * } \) as \( t \rightarrow \infty \) . Then for some \( \varepsilon > 0 \) , there exist points \( {t}_{i} \) tending to \( + \infty \) such that \( \begin{Vmatrix}{x\left( {t}_{i}\right) - {x}^{ * }}\end{Vmatrix} \geq \varepsilon \left( {i = 1,2,\ldots }\right) \) . We can assume \( {t}_{i + 1} - {t}_{i} > \varepsilon /\left( {2k}\right) \) . Let \( \eta > 0 \) be such that \[ {\begin{Vmatrix}x - {x}^{ * }\end{Vmatrix}}_{\infty } \geq \begin{Vmatrix}{u - {x}^{ * }}\end{Vmatrix} \geq \frac{\varepsilon }{2} \Longrightarrow W\left( u\right) \geq \eta \] (such \( \eta \) exists because \( W \) is continuous and positive on the (nonempty) annulus in question). Then \[ \left| {t - {t}_{i}}\right| < \frac{\varepsilon }{2k} \Longrightarrow \begin{Vmatrix}{x\left( t\right) - x\left( {t}_{i}\right) }\end{Vmatrix} < \frac{\varepsilon }{2} \Longrightarrow \begin{Vmatrix}{x\left( t\right) - {x}^{ * }}\end{Vmatrix} \geq \frac{\varepsilon }{2}, \] so that \[ {\int }_{{t}_{i - 1}}^{{t}_{i + 1}}W\left( {x\left( \tau \right) }\right) {d\tau } \geq \frac{\eta \varepsilon }{k} \] This would imply that \( {\int }_{0}^{t}W\left( {x\left( \tau \right) }\right) {d\tau } \) diverges, a contradiction that completes the proof. ## Construction of a Stabilizing Feedback The proof of Theorem 5.5 just given says nothing explicitly about how to construct a trajectory converging to \( {x}^{ * } \), but in fact this is implicit from the constructive nature in which we proved Theorem 2.4, which lies at the heart of Theorem 5.7 and hence of Theorem 5.5 itself. Let us now make explicit the proximal aiming method underlying these results. We define a function \( f : {\mathbb{R}}^{n} \times \mathbb{R} \times \mathbb{R} \rightarrow {\mathbb{R}}^{n} \) as follows: given \( \left( {x, y, r}\right) \in \) \( {\mathbb{R}}^{n} \times \mathbb{R} \times \mathbb{R} \), select any \( \left( {{x}^{\prime },{y}^{\prime },{r}^{\prime }}\right) \) in \( {\operatorname{proj}}_{S}\left( {x, y, r}\right) \), where \[ S \mathrel{\text{:=}} \left\{ {\left( {{x}^{\prime },{y}^{\prime },{r}^{\prime }}\right) : Q\left( {x}^{\prime }\right) + {y}^{\prime } \leq {r}^{\prime }}\right\} . \] (This is the set used in the proof of Theorem 5.7 when \( \varphi = Q\left( x\right) + y \), as is the case in the proof of Theorem 5.5.) Now select any \( v \in F\left( {x}^{\prime }\right) \) which minimizes over \( F\left( {x}^{\prime }\right) \) the function \( v \mapsto \) \( \left\langle {v, x - {x}^{\prime }}\r
1012_(GTM178)Nonsmooth Analysis and Control Theory
76
at the heart of Theorem 5.7 and hence of Theorem 5.5 itself. Let us now make explicit the proximal aiming method underlying these results. We define a function \( f : {\mathbb{R}}^{n} \times \mathbb{R} \times \mathbb{R} \rightarrow {\mathbb{R}}^{n} \) as follows: given \( \left( {x, y, r}\right) \in \) \( {\mathbb{R}}^{n} \times \mathbb{R} \times \mathbb{R} \), select any \( \left( {{x}^{\prime },{y}^{\prime },{r}^{\prime }}\right) \) in \( {\operatorname{proj}}_{S}\left( {x, y, r}\right) \), where \[ S \mathrel{\text{:=}} \left\{ {\left( {{x}^{\prime },{y}^{\prime },{r}^{\prime }}\right) : Q\left( {x}^{\prime }\right) + {y}^{\prime } \leq {r}^{\prime }}\right\} . \] (This is the set used in the proof of Theorem 5.7 when \( \varphi = Q\left( x\right) + y \), as is the case in the proof of Theorem 5.5.) Now select any \( v \in F\left( {x}^{\prime }\right) \) which minimizes over \( F\left( {x}^{\prime }\right) \) the function \( v \mapsto \) \( \left\langle {v, x - {x}^{\prime }}\right\rangle \), and set \( f\left( {x, y, r}\right) = v \) . The following restates Corollary 2.6 in the present setting, where the initial condition on \( \left( {x, y, r}\right) \) is taken to be \( \left( {\alpha ,0, Q\left( \alpha \right) }\right) \in S \) ; see also Exercise 1.14. 5.8. Proposition. Let \( \left( {Q, W}\right) \) be a Lyapounov pair for \( {x}^{ * } \), where \( W \) has the additional properties of Exercise 5.4. Then, for any \( \alpha \in \operatorname{dom}Q \), there exists at least one Euler solution \( \left( {x, y}\right) \) on \( \lbrack 0,\infty ) \) of the initial-value problem \[ \dot{x}\left( t\right) = f\left( {x\left( t\right), y\left( t\right), Q\left( \alpha \right) }\right) ,\;\dot{y}\left( t\right) = W\left( {x\left( t\right) }\right) ,\;x\left( 0\right) = \alpha, y\left( 0\right) = 0, \] and any such solution defines a trajectory \( x\left( \cdot \right) \) of \( F \) on \( \lbrack 0,\infty ) \) which converges to \( {x}^{ * } \) as \( t \rightarrow \infty \) . The function \( f \) above fails to satisfy \( f\left( {x, y, r}\right) \in F\left( x\right) \) in general; it is not a feedback selection for \( F \) . When \( F \) is locally Lipschitz however, we showed in Theorem 3.4 how to define a selection \( g \) that can play the role of \( f \) . In our present setting, we define \( g\left( {x, y, r}\right) \) as follows: given \( f\left( {x, y, r}\right) = v \in F\left( {x}^{\prime }\right) \) as above, let \( w \) be the point in \( F\left( x\right) \) closest to \( v \) ; set \( g\left( {x, y, r}\right) = w \in F\left( x\right) \) . Then we have 5.9. Proposition. In the context of Proposition 5.8, when \( F \) is locally Lipschitz, there is at least one Euler solution \( \left( {x, y}\right) \) to the initial-value problem \[ \dot{x}\left( t\right) = g\left( {x\left( t\right), y\left( t\right), Q\left( \alpha \right) }\right) ,\;\dot{y}\left( t\right) = W\left( {x\left( t\right) }\right) ,\;x\left( 0\right) = \alpha, y\left( 0\right) = 0, \] and any such solution defines a trajectory \( x \) of \( F \) on \( \lbrack 0,\infty ) \) which converges to \( {x}^{ * } \) as \( t \rightarrow \infty \) . The construction of the "stabilizing feedback" in Proposition 5.9 is quite explicit. Note however that it requires the calculation of another variable \( y\left( t\right) = {\int }_{0}^{t}W\left( {x\left( \tau \right) }\right) {d\tau } \) involving the past history of \( x \) ; this is an instance of dynamic feedback. It is of considerable interest to be able to define a static feedback selection \( g\left( x\right) \in F\left( x\right) \) depending only on \( x \), and giving rise to trajectories converging to the equilibrium, whatever the initial condition. When a Lyapounov pair \( \left( {Q, W}\right) \) exists having \( Q \) locally Lipschitz, this can be done by proximal aiming techniques, but doing so would take us beyond the scope of this chapter. ## 6 Monotonicity and Attainability We have seen the relevance of weak decrease along trajectories in \( §5 \), in connection with asymptotic controllability. This and other monotonicity issues will arise again in other contexts, and we will require some corresponding localized proximal characterizations. We will derive these first in the autonomous setting. Let \( \Omega \) be an open subset of \( {\mathbb{R}}^{n} \), and let \( \varphi \in \mathcal{F}\left( \Omega \right) \) . Extending our earlier definition (which dealt with the case \( \Omega = {\mathbb{R}}^{n} \) ), we say that \( \left( {\varphi, F}\right) \) is weakly decreasing on \( \Omega \) if for any \( \alpha \in \operatorname{dom}\varphi \subset \Omega \), there exists a trajectory \( x \) of \( F \) on \( \lbrack 0,\infty ) \) with \( x\left( 0\right) = \alpha \) having the property that for any interval \( \left\lbrack {0, T}\right\rbrack \) for which \( x\left( \left\lbrack {0, T}\right\rbrack \right) \subset \Omega \), we have \[ \varphi \left( {x\left( t\right) }\right) \leq \varphi \left( \alpha \right) \forall t \in \left\lbrack {0, T}\right\rbrack \] Recall that trajectories \( x \) on \( \left\lbrack {0, T}\right\rbrack \) can always be extended to \( \lbrack 0,\infty ) \), but above we only require \( \varphi \left( {x\left( t\right) }\right) \leq \varphi \left( \alpha \right) \) to hold until the exit time \[ \tau \left( {x,\Omega }\right) \mathrel{\text{:=}} \inf \{ t \geq 0 : x\left( t\right) \in \operatorname{comp}\Omega \} . \] Of course, \( \tau \) can equal \( + \infty \), precisely when \( x\left( t\right) \in \Omega \forall t \geq 0 \) . 6.1. Theorem. Let \( \varphi \in \mathcal{F}\left( \Omega \right) \) . The system \( \left( {\varphi, F}\right) \) is weakly decreasing on \( \Omega \) iff \[ h\left( {x,{\partial }_{P}\varphi \left( x\right) }\right) \leq 0\forall x \in \Omega . \] Proof. Let \( \left( {\varphi, F}\right) \) be weakly decreasing on \( \Omega \), and let \( \alpha \in \Omega \) and \( \zeta \in {\partial }_{P}\varphi \left( \alpha \right) \) be given. Let \( \delta > 0 \) be such that \( \bar{B}\left( {\alpha ;\delta }\right) \subset \Omega \), and define \[ S \mathrel{\text{:=}} \left\{ {\left( {x, r}\right) \in {\mathbb{R}}^{n} \times \mathbb{R} : x \in \bar{B}\left( {\alpha ;\delta }\right) ,\varphi \left( x\right) \leq r}\right\} . \] Then \[ \left( {\zeta , - 1}\right) \in {N}_{S}^{P}\left( {\alpha ,\varphi \left( \alpha \right) }\right) \] Also define \[ \widetilde{F}\left( {x, r}\right) \mathrel{\text{:=}} F\left( x\right) \times \{ 0\} \text{ if }x \in B\left( {\alpha ;\delta }\right) , \] and otherwise set \[ \widetilde{F}\left( {x, r}\right) \mathrel{\text{:=}} \operatorname{co}\left\{ {\mathop{\bigcup }\limits_{{\parallel y - \alpha \parallel = \delta }}F\left( y\right) \cup \{ 0\} }\right\} \times \{ 0\} . \] Then \( \widetilde{F} \) satisfies the Standing Hypotheses and \( \left( {S,\widetilde{F}}\right) \) is weakly invariant (why?). It follows from Theorem 2.10 that for some \( \left( {v,0}\right) \in \widetilde{F}\left( {\alpha ,\varphi \left( \alpha \right) }\right) \) we have \( \langle \left( {v,0}\right) ,\left( {\zeta , - 1}\right) \rangle \leq 0 \) . This implies \( h\left( {\alpha ,\zeta }\right) \leq 0 \), whence the Hamiltonian inequality holds everywhere in \( \Omega \) . For the converse, let the Hamiltonian inequality hold, and let \( \alpha \in \operatorname{dom}\varphi \subset \) \( \Omega \) be given. Let \( {\Omega }_{k} \mathrel{\text{:=}} \Omega \cap B\left( {\alpha ;k}\right) \), and define a closed set \( {S}_{k} \) and a multifunction \( {\widetilde{F}}_{k} \) by \[ {S}_{k} \mathrel{\text{:=}} \left\{ {\left( {x, r}\right) \in {\mathbb{R}}^{n} \times \mathbb{R} : x \in {\Omega }_{k},\varphi \left( x\right) \leq r}\right\} \cup \left( {\operatorname{comp}{\Omega }_{k} \times \mathbb{R}}\right) , \] \[ {\widetilde{F}}_{k}\left( {x, r}\right) \mathrel{\text{:=}} F\left( x\right) \times \{ 0\} \;\text{ if }x \in {\Omega }_{k}, \] and otherwise \[ {\widetilde{F}}_{k}\left( {x, r}\right) \mathrel{\text{:=}} \operatorname{co}\left\{ {\mathop{\bigcup }\limits_{{y \in \text{ bdry }{\Omega }_{k}}}F\left( x\right) \cup \{ 0\} }\right\} \times \{ 0\} . \] We claim that \( \left( {{S}_{k},{\widetilde{F}}_{k}}\right) \) is weakly invariant. Let us verify this by the criterion of Theorem 2.4. If \( x \notin {\Omega }_{k} \), it is evidently satisfied, since \( 0 \in {\widetilde{F}}_{k}\left( {x, r}\right) \) . If \( x \in \) \( {\Omega }_{k} \), then a proximal normal \( \left( {\zeta ,\lambda }\right) \) to \( {S}_{k} \) at \( \left( {x, r}\right) \) belongs to \( {N}_{\text{epi }\varphi }^{P}\left( {x, r}\right) \), and precisely the argument used to prove the "converse" part of Theorem 5.7 demonstrates that \( \widetilde{h}\left( {x, r,\zeta ,\lambda }\right) \leq 0 \) . This establishes the claim. Since \( \left( {{S}_{1},{\widetilde{F}}_{1}}\right) \) is weakly invariant, we deduce the existence of a trajectory \( x \) for \( F \) on \( \left\lbrack {0, T}\right\rbrack \) with \( x\left( 0\right) = \alpha \) such that \( \varphi \left( {x\left( t\right) }\right) \leq \varphi \left( \alpha \right) \) for \( t \in \left\lbrack {0,{\tau }_{1}}\right) \) , where \( {\tau }_{1} \) is the exit time of \( x \) from \( {\Omega }_{1} \) . If \( {\tau }_{1} = \infty \), or if \( x\left( {\tau }_{1}\right) \in \operatorname{bdry}\Omega \), then the trajectory \( x \) satisfies the requirement of the definition whereby \( \left( {\varphi, F}\right) \) is weakly decreasing on \( \Omega \) . Otherwise, \( x\left( {\tau }_{1}\right) \in \Omega \) and \( \begin{Vmatrix}{x\left( {\tau }_{1}\right) - \alpha }\end{Vmatrix} = 1 \) . In this case, we invoke the weak invariance of \( \left( {{S}_{2},{\widetilde{F}}_{2}}\right) \) to construct a trajectory that extends \( x \), beginning at \( x\left( {\tau }_{1}\right) \), to the interval \( \left\lbrack {0,{\tau }_{1} + {\tau }_{2}}\right) \), where \( {\ta
1012_(GTM178)Nonsmooth Analysis and Control Theory
77
ctory \( x \) for \( F \) on \( \left\lbrack {0, T}\right\rbrack \) with \( x\left( 0\right) = \alpha \) such that \( \varphi \left( {x\left( t\right) }\right) \leq \varphi \left( \alpha \right) \) for \( t \in \left\lbrack {0,{\tau }_{1}}\right) \) , where \( {\tau }_{1} \) is the exit time of \( x \) from \( {\Omega }_{1} \) . If \( {\tau }_{1} = \infty \), or if \( x\left( {\tau }_{1}\right) \in \operatorname{bdry}\Omega \), then the trajectory \( x \) satisfies the requirement of the definition whereby \( \left( {\varphi, F}\right) \) is weakly decreasing on \( \Omega \) . Otherwise, \( x\left( {\tau }_{1}\right) \in \Omega \) and \( \begin{Vmatrix}{x\left( {\tau }_{1}\right) - \alpha }\end{Vmatrix} = 1 \) . In this case, we invoke the weak invariance of \( \left( {{S}_{2},{\widetilde{F}}_{2}}\right) \) to construct a trajectory that extends \( x \), beginning at \( x\left( {\tau }_{1}\right) \), to the interval \( \left\lbrack {0,{\tau }_{1} + {\tau }_{2}}\right) \), where \( {\tau }_{2} \) is the exit time of the extension from \( {\Omega }_{2} \), and where for \( t \in \left\lbrack {{\tau }_{1},{\tau }_{1} + {\tau }_{2}}\right) \) we have \( \varphi \left( {x\left( t\right) }\right) \leq \varphi \left( {x\left( {\tau }_{1}\right) }\right) \leq \varphi \left( \alpha \right) \) . Again, if either \( {\tau }_{2} = \infty \) or \( x\left( {{\tau }_{1} + {\tau }_{2}}\right) \in \operatorname{bdry}\Omega \) , we have the required trajectory; otherwise \( x\left( {{\tau }_{1} + {\tau }_{2}}\right) \in \Omega ,\begin{Vmatrix}{x\left( {{\tau }_{1} + {\tau }_{2}}\right) - \alpha }\end{Vmatrix} = \) 2, and we begin again. If the process fails to end after finitely many steps, the resulting trajectory \( x \) is defined on \( \lbrack 0,\infty ) \), since \[ \begin{Vmatrix}{x\left( {\mathop{\sum }\limits_{1}^{k}{\tau }_{i}}\right) - \alpha }\end{Vmatrix} = k\text{ implies that }\mathop{\sum }\limits_{1}^{k}{\tau }_{i} \rightarrow \infty \text{ as }k \rightarrow \infty \] (i.e., trajectories of \( F \) do not blow up in finite time). Then \( x \) on \( \lbrack 0,\infty ) \) is itself the required trajectory, and the proof is complete. ## Strongly Decreasing Systems on \( \Omega \) Let \( \varphi \in \mathcal{F}\left( \Omega \right) \), where \( \Omega \) is an open subset of \( {\mathbb{R}}^{n} \) . The system \( \left( {\varphi, F}\right) \) is said to be strongly decreasing on \( \Omega \) if for any \( \alpha \in \operatorname{dom}\varphi \subset \Omega \), for any trajectory \( x \) of \( F \) on an interval \( \left\lbrack {a, b}\right\rbrack \) which lies in \( \Omega \) and satisfies \( x\left( a\right) = \alpha \), we have \[ \varphi \left( {x\left( t\right) }\right) \leq \varphi \left( \alpha \right) \forall t \in \left\lbrack {a, b}\right\rbrack . \] 6.2. Exercise. Show that \( \left( {\varphi, F}\right) \) is strongly decreasing on \( \Omega \) iff every trajectory \( x \) of \( F \) on an interval \( \left\lbrack {a, b}\right\rbrack \) which lies in \( \Omega \) is such that the function \( t \mapsto \varphi \left( {x\left( t\right) }\right) \) is decreasing on \( \left\lbrack {a, b}\right\rbrack \) . 6.3. Theorem. Let \( F \) be locally Lipschitz. Then \( \left( {\varphi, F}\right) \) is strongly decreasing on \( \Omega \) iff \[ H\left( {x,{\partial }_{P}\varphi \left( x\right) }\right) \leq 0\forall x \in \Omega . \] Proof. Let \( \left( {\varphi, F}\right) \) be strongly decreasing on \( \Omega \), and let \( \zeta \in {\partial }_{P}\varphi \left( \alpha \right) \), where \( \alpha \in \Omega \) . Fix any \( {v}_{0} \in F\left( \alpha \right) \) . We wish to prove that \( \left\langle {{v}_{0},\zeta }\right\rangle \leq 0 \), to allow us to deduce the Hamiltonian inequality of the theorem. Pick \( \delta > 0 \) such that \( \bar{B}\left( {\alpha ;\delta }\right) \subset \Omega \), and set \[ S \mathrel{\text{:=}} \left\{ {\left( {x, r}\right) \in {\mathbb{R}}^{n} \times \mathbb{R} : x \in \bar{B}\left( {\alpha ;\delta }\right) ,\varphi \left( x\right) \leq r}\right\} . \] Let \( f\left( x\right) \mathrel{\text{:=}} v \) in \( F\left( x\right) \) closest to \( {v}_{0} \) . Then \( f \) is continuous (Exercise 3.3(c)) and \( f\left( \alpha \right) = {v}_{0} \) . We define the multifunction \[ \widetilde{F}\left( {x, r}\right) \mathrel{\text{:=}} \{ \left( {f\left( x\right) ,0}\right) \} \;\text{ if }x \in B\left( {\alpha ;\delta }\right) , \] \[ \mathrel{\text{:=}} \operatorname{co}\{ 0, f\left( y\right) : \parallel y - \alpha \parallel = \delta \} \times \{ 0\} \;\text{ otherwise. } \] Then \( S \) is closed, \( \widetilde{F} \) satisfies the Standing Hypotheses, and \( \left( {S,\widetilde{F}}\right) \) is weakly invariant. Since \( \zeta \in {\partial }_{P}\varphi \left( \alpha \right) \), we have \( \left( {\zeta , - 1}\right) \in {N}_{S}^{P}\left( {\alpha ,\varphi \left( \alpha \right) }\right) \), so that by Theorem 2.10, \[ \langle \left( {f\left( \alpha \right) ,0}\right) ,\left( {\zeta , - 1}\right) \rangle = \left\langle {{v}_{0},\zeta }\right\rangle \leq 0 \] proving the necessity. For the converse, let the Hamiltonian inequality hold, and let \( \widehat{x} \) be an Euler arc on \( \left\lbrack {0, T}\right\rbrack \) lying in \( \Omega \) . In order to prove that \( \varphi \left( {\widehat{x}\left( t\right) }\right) \) is decreasing, it suffices to prove that for any \( a \in \lbrack 0, T) \), for all \( b \in (a, T\rbrack \) sufficiently close to \( a \), we have \( \varphi \left( {\widehat{x}\left( b\right) }\right) \leq \varphi \left( {\widehat{x}\left( a\right) }\right) \) (we may assume \( \varphi \left( {\widehat{x}\left( a\right) }\right) \) finite). To this end, set \[ S \mathrel{\text{:=}} \operatorname{cl}\{ \left( {x, r}\right) : x \in \Omega, r \geq \varphi \left( x\right) \} \] and pick \( M > 0 \) such that \( \widehat{x}\left( a\right) + {4M}\bar{B} \subset \Omega \), and then pick \( b > a \) sufficiently near \( a \) so that \[ \parallel \widehat{x}\left( t\right) - \widehat{x}\left( a\right) \parallel < M\forall t \in \left\lbrack {a, b}\right\rbrack . \] It follows that for any \[ \left( {x, r}\right) \in \left( {\widehat{x}\left( a\right) + M\bar{B}}\right) \times \left( {\varphi \left( {\widehat{x}\left( a\right) }\right) + M\bar{B}}\right) , \] for any \( \left( {{x}^{\prime },{r}^{\prime }}\right) \in {\operatorname{proj}}_{S}\left( {x, r}\right) \), we have \[ \begin{Vmatrix}{\left( {{x}^{\prime },{r}^{\prime }}\right) - \left( {\widehat{x}\left( a\right) ,\varphi \left( {\widehat{x}\left( a\right) }\right) }\right) }\end{Vmatrix} < {4M}. \] Now let \( f \) be a selection of \( F \) whose unique Euler arc beginning at \( \left( {a,\widehat{x}\left( a\right) }\right) \) is \( \widehat{x} \) (Corollary 3.7). If \[ \left( {x, r}\right) \in \left( {\widehat{x}\left( a\right) + M\bar{B}}\right) \times \left( {\varphi \left( {\widehat{x}\left( a\right) }\right) + M\bar{B}}\right) , \] then \[ \left( {\zeta ,\lambda }\right) \mathrel{\text{:=}} \left( {x - {x}^{\prime }, r - {r}^{\prime }}\right) \in {N}_{S}^{P}\left( {{x}^{\prime },{r}^{\prime }}\right) , \] where \( {x}^{\prime } \in \widehat{x}\left( a\right) + {4M}\bar{B} \subset \Omega \) . It follows that \( \left( {\zeta ,\lambda }\right) \in {N}_{\text{epi }\varphi }^{P}\left( {{x}^{\prime },{r}^{\prime }}\right) \) . If \( \lambda < 0 \) , then \( - \zeta /\lambda \in {\partial }_{P}\varphi \left( {x}^{\prime }\right) \), so by hypothesis we have \( \left\langle {f\left( {x}^{\prime }\right) , - \zeta /\lambda }\right\rangle \leq 0 \), whence \( \left\langle {f\left( {x}^{\prime }\right), x - {x}^{\prime }}\right\rangle \leq 0 \) . If \( \lambda = 0 \), then \[ \left( {\zeta ,0}\right) \in {N}_{\text{epi }\varphi }^{P}\left( {\widehat{x}\left( a\right) ,\varphi \left( {\widehat{x}\left( a\right) }\right) }\right) , \] and Problem 1.11.23 implies the existence of \[ {x}_{i}\overset{\varphi }{ \rightarrow }{x}^{\prime },\;{\varepsilon }_{i} \rightarrow 0\left( {\text{with }{\varepsilon }_{i} > 0}\right) ,\;{\zeta }_{i} \rightarrow \zeta , \] such that \( \left( {{\zeta }_{i}, - {\varepsilon }_{i}}\right) \in {\partial }_{P}\varphi \left( {x}_{i}\right) \) . Since \( F \) is Lipschitz, there exist \( {v}_{i} \in F\left( {x}_{i}\right) \) with \( \begin{Vmatrix}{{v}_{i} - f\left( {x}^{\prime }\right) }\end{Vmatrix} \leq K\begin{Vmatrix}{{x}_{i} - {x}^{\prime }}\end{Vmatrix} \) . As above, we have \( \left\langle {{\zeta }_{i},{v}_{i}}\right\rangle \leq 0 \), and a passage to the limit gives \( \left\langle {\zeta, f\left( {x}^{\prime }\right) }\right\rangle \leq 0 \) once more. The above confirms that Proposition 2.1 applies to \( S \) together with the map \( \left( {x, r}\right) \mapsto \left( {f\left( x\right) ,0}\right) \) . Since the unique Euler arc for this map beginning at \( \left( {\widehat{x}\left( a\right) ,\varphi \left( {\widehat{x}\left( a\right) }\right) }\right) \) is \( \left( {\widehat{x}\left( \cdot \right) ,\varphi \left( {\widehat{x}\left( a\right) }\right) }\right) \), we deduce \[ {d}_{S}\left( {\widehat{x}\left( t\right) ,\varphi \left( {\widehat{x}\left( a\right) }\right) }\right) \leq {d}_{S}\left( {\widehat{x}\left( a\right) ,\varphi \left( {\widehat{x}\left( a\right) }\right) }\right) = 0\forall t \in \left\lbrack {a, b}\right\rbrack . \] This implies \( \varphi \left( {\widehat{x}\left( b\right) }\right) \leq \varphi \left( {\widehat{x}\left( a\right) }\right) \), as required. ## A Nonautonomous Extension Suppose now that \( \varphi \) and \( F \) depend on \( t \) as well as \( x \) . We can easily extend the monotonicity characterizations obtained above to this nonautonomous case, and make them relative to a given \( t \) -interval \( \left( {{t}_{0},{t}_{1}}\right) \), where we allow \( {t}_{0} = - \infty \) and/or \( {t}_{1} = \infty \) . We will say that \( \left( {\varphi, F}\right) \) is weakly decreasing on \( \left( {{t}_{0},{t}_{1}}\right)
1012_(GTM178)Nonsmooth Analysis and Control Theory
78
at{x}\left( \cdot \right) ,\varphi \left( {\widehat{x}\left( a\right) }\right) }\right) \), we deduce \[ {d}_{S}\left( {\widehat{x}\left( t\right) ,\varphi \left( {\widehat{x}\left( a\right) }\right) }\right) \leq {d}_{S}\left( {\widehat{x}\left( a\right) ,\varphi \left( {\widehat{x}\left( a\right) }\right) }\right) = 0\forall t \in \left\lbrack {a, b}\right\rbrack . \] This implies \( \varphi \left( {\widehat{x}\left( b\right) }\right) \leq \varphi \left( {\widehat{x}\left( a\right) }\right) \), as required. ## A Nonautonomous Extension Suppose now that \( \varphi \) and \( F \) depend on \( t \) as well as \( x \) . We can easily extend the monotonicity characterizations obtained above to this nonautonomous case, and make them relative to a given \( t \) -interval \( \left( {{t}_{0},{t}_{1}}\right) \), where we allow \( {t}_{0} = - \infty \) and/or \( {t}_{1} = \infty \) . We will say that \( \left( {\varphi, F}\right) \) is weakly decreasing on \( \left( {{t}_{0},{t}_{1}}\right) \times \Omega \) if for any \( \tau \in \left( {{t}_{0},{t}_{1}}\right) \) and \( \alpha \in \Omega \) there exists a trajectory \( x \) on \( \left\lbrack {\tau ,{t}_{1}}\right) \) with \( x\left( \tau \right) = \alpha \) such that \[ \varphi \left( {t, x\left( t\right) }\right) \leq \varphi \left( {\tau ,\alpha }\right) \forall t \in \left\lbrack {\tau, b}\right\rbrack \] where \( \left\lbrack {\tau, b}\right\rbrack \) is any subinterval of \( \left\lbrack {\tau ,{t}_{1}}\right) \) upon which \( x\left( t\right) \) remains in \( \Omega \) . If this holds for all trajectories \( x \), then \( \left( {\varphi, F}\right) \) is said to be strongly decreasing on \( \left( {{t}_{0},{t}_{1}}\right) \times \Omega \) . The state augmentation device introduced in \( §3 \) , together with Theorems 6.1 and 6.3 , leads easily to: 6.4. Exercise. Let \( \varphi \in \mathcal{F}\left( {\left( {{t}_{0},{t}_{1}}\right) \times \Omega }\right) \) . (a) Then \( \left( {\varphi, F}\right) \) is weakly decreasing on \( \left( {{t}_{0},{t}_{1}}\right) \times \Omega \) iff \[ \theta + h\left( {t, x,\zeta }\right) \leq 0\forall \left( {\theta ,\zeta }\right) \in {\partial }_{P}\varphi \left( {t, x}\right) ,\forall \left( {t, x}\right) \in \left( {{t}_{0},{t}_{1}}\right) \times \Omega . \] (b) If \( F \) is locally Lipschitz, then \( \left( {\varphi, F}\right) \) is strongly decreasing on \( \left( {{t}_{0},{t}_{1}}\right) \times \Omega \) iff \[ \theta + H\left( {t, x,\zeta }\right) \leq 0\forall \left( {\theta ,\zeta }\right) \in {\partial }_{P}\varphi \left( {t, x}\right) ,\forall \left( {t, x}\right) \in \left( {{t}_{0},{t}_{1}}\right) \times \Omega . \] A certain number of other monotonicity properties of systems have not been discussed. We will conclude with strong increase, leaving to the end-of-chapter problems an exhaustive study of all the variations on this theme. In the nonautonomous case still, \( \left( {\varphi, F}\right) \) is said to be strongly increasing on \( \left( {{t}_{0},{t}_{1}}\right) \times \Omega \) provided that for any interval \( \left\lbrack {a, b}\right\rbrack \) contained in \( \left( {{t}_{0},{t}_{1}}\right) \), for any trajectory \( x \) of \( F \) on \( \left\lbrack {a, b}\right\rbrack \) for which \( x\left( t\right) \in \Omega \forall t \in \left\lbrack {a, b}\right\rbrack \), we have \[ \varphi \left( {t, x\left( t\right) }\right) \leq \varphi \left( {b, x\left( b\right) }\right) \forall t \in \left\lbrack {a, b}\right\rbrack . \] Of course, this last inequality is automatically satisfied when \( \left( {b, x\left( b\right) }\right) \notin \) \( \operatorname{dom}\varphi \) . As in the autonomous case, this strong increase property is equivalent to the requirement that the function \( t \mapsto \varphi \left( {t, x\left( t\right) }\right) \) be increasing on \( \left\lbrack {a, b}\right\rbrack \) whenever \( x \) is a trajectory on some interval \( \left\lbrack {a, b}\right\rbrack \subset \left( {{t}_{0},{t}_{1}}\right) \) for which \( x\left( t\right) \) remains in \( \Omega \) . 6.5. Proposition. Let \( F \) be locally Lipschitz. Then \( \left( {\varphi, F}\right) \) is strongly increasing on \( \left( {{t}_{0},{t}_{1}}\right) \times \Omega \) iff \[ \theta + h\left( {t, x,\zeta }\right) \geq 0\forall \left( {\theta ,\zeta }\right) \in {\partial }_{P}\varphi \left( {t, x}\right) ,\forall \left( {t, x}\right) \in \left( {{t}_{0},{t}_{1}}\right) \times \Omega . \] Proof. Let \( x \) be a trajectory of \( F \) on \( \left( {{t}_{0},{t}_{1}}\right) \), and define \( y\left( t\right) \mathrel{\text{:=}} x\left( {{t}^{ * } - t}\right) \) , where \( {t}^{ * } \) is a point in \( \left( {{t}_{0},{t}_{1}}\right) \) . Then the function \( y \) is defined on the interval \( \left( {{t}^{ * } - {t}_{1},{t}^{ * } - {t}_{0}}\right) \), and we have \[ \dot{y}\left( t\right) = - \dot{x}\left( {{t}^{ * } - t}\right) \in - F\left( {{t}^{ * } - t, x\left( {{t}^{ * } - t}\right) }\right) = - F\left( {{t}^{ * } - t, y\left( t\right) }\right) \text{ a.e. } \] This shows that \( y \) is a trajectory of \( {F}^{ * } \) on \( \left( {{t}^{ * } - {t}_{1},{t}^{ * } - {t}_{0}}\right) \), where \( {F}^{ * }\left( {t, y}\right) \mathrel{\text{:=}} \) \( - F\left( {{t}^{ * } - t, y}\right) \) . Clearly this correspondence between \( x \) and \( y \) is one-to-one, so that \( \varphi \left( {t, x\left( t\right) }\right) \) is increasing for all such \( x \) (while \( x\left( t\right) \in \Omega \) ) iff \( \varphi \left( {{t}^{ * } - t, y\left( t\right) }\right) \) is decreasing for all such \( y \) (in each case on the relevant interval). Letting \( {\varphi }^{ * }\left( {t, y}\right) \mathrel{\text{:=}} \varphi \left( {{t}^{ * } - t, y}\right) \), we have proven that strong increase on \( \left( {{t}_{0},{t}_{1}}\right) \times \Omega \) for \( \left( {\varphi, F}\right) \) is the same as strong decrease on \( \left( {{t}^{ * } - {t}_{1},{t}^{ * } - {t}_{0}}\right) \times \Omega \) for \( \left( {{\varphi }^{ * },{F}^{ * }}\right) \) . Applying Exercise 6.4(b), and denoting by \( {H}^{ * } \) the upper Hamiltonian of \( {F}^{ * } \), the latter is equivalent to \[ \theta + {H}^{ * }\left( {t, y,\zeta }\right) \leq 0\;\forall \left( {\theta ,\zeta }\right) \in {\partial }_{P}{\varphi }^{ * }\left( {t, y}\right) ,\;\forall \left( {t, y}\right) \in \left( {{t}^{ * } - {t}_{1},{t}^{ * } - {t}_{0}}\right) \times \Omega . \] But it is easy to see that \[ {H}^{ * }\left( {t, y,\zeta }\right) = - h\left( {{t}^{ * } - t, y,\zeta }\right) \] \[ \left( {\theta ,\zeta }\right) \in {\partial }_{P}{\varphi }^{ * }\left( {t, y}\right) \Leftrightarrow \left( {-\theta ,\zeta }\right) \in {\partial }_{P}\varphi \left( {{t}^{ * } - t, y}\right) . \] The proposition then follows immediately. ## Local Attainability We have addressed the issue of remaining in a given set, and also that of asymptotically approaching an equilibrium. We now examine under what conditions an initial-value outside a given set can be steered to it in finite time. We say that the system \( \left( {S, F}\right) \) is locally attainable if there exists \( r > 0 \) and \( T > 0 \) such that the following holds: for all \( \alpha \) having \( {d}_{S}\left( \alpha \right) < r \), there exists a trajectory \( x \) of \( F \) on \( \lbrack 0,\infty ) \) such that \[ x\left( 0\right) = \alpha \text{and}x\left( t\right) \in S\forall t \geq T\text{.} \] 6.6. Theorem. Let \( S \) be compact, and let \( F \) be locally Lipschitz. Suppose that for some \( \delta > 0 \) we have \[ h\left( {x,\zeta }\right) \leq - \delta \parallel \zeta \parallel \forall \zeta \in {N}_{S}^{P}\left( x\right) ,\forall x \in S. \] (1) Then \( \left( {S, F}\right) \) is locally attainable. Proof. Let \( F \) be Lipschitz of rank \( K \) on \( S + {\eta B} \), where \( \eta > 0 \), and let \( r > 0 \) satisfy \( r < \min \left( {\delta /K,\eta }\right) \) . Set \( \lambda \mathrel{\text{:=}} \delta - {Kr} \) . We claim that for \( y \in \mathbb{R} \), the system \( \left( {{d}_{S}\left( x\right) + {\lambda y}, F\left( x\right) \times \{ 1\} }\right) \) is weakly decreasing on \[ \Omega \mathrel{\text{:=}} \{ \left( {S + {rB}}\right) \smallsetminus S\} \times \mathbb{R}. \] We will verify this by means of the criterion of Theorem 6.1. Any proximal subgradient \( \left( {\zeta ,\theta }\right) \) of \( {d}_{S}\left( x\right) + {\lambda y} \) in \( \Omega \) is of the form \( \left( {\zeta ,\lambda }\right) \) where \( \zeta \in {\partial }_{P}{d}_{S}\left( x\right) \) . According to Theorem 1.6.1, \( \parallel \zeta \parallel = 1 \) and \( \zeta \in {N}_{S}^{P}\left( s\right) \), where \( s \in {\operatorname{proj}}_{S}\left( x\right) \) . By hypothesis (1), we have \( h\left( {s,\zeta }\right) + \delta \leq 0 \) . The Lipschitz condition on \( F \) gives \[ h\left( {x,\zeta }\right) + \lambda \leq h\left( {s,\zeta }\right) + K\parallel x - s\parallel + \delta - {Kr} \] \[ \leq - \delta + K{d}_{S}\left( x\right) + \delta - {Kr} < 0. \] This proves the claim. It follows that for any \( \alpha \in \left( {S + {rB}}\right) \smallsetminus S \), there is a trajectory \( x \) of \( F \) on \( \lbrack 0,\infty ) \) such that \( x\left( 0\right) = \alpha \), and such that \[ {d}_{S}\left( {x\left( t\right) }\right) + {\lambda t} \leq {d}_{S}\left( \alpha \right) \] for all \( t > 0 \) up to the first \( T > 0 \) such that \( {d}_{S}\left( {x\left( T\right) }\right) = 0 \) ; i.e., such that \( x\left( T\right) \in S \) . Note that since \( \lambda > 0 \), such \( T \) must exist (uniformly for \( \alpha \) ). Once \( x\left( T\right) \in S \), then there is a trajectory beginning at \( x\left( T\right) \) which extends \( x \) and remains in \( S \) thereafter, since (1) implies that \( \left( {S, F}\right) \) is weakly invariant. 6.7. Exercise. (a) Prove that in Theorem 6.6, the parameter \( T \) can be taken to be \( r/\left( {\delta - {Kr}}\right) \) . (b) The time required to steer \(
1012_(GTM178)Nonsmooth Analysis and Control Theory
79
ht) + \delta - {Kr} < 0. \] This proves the claim. It follows that for any \( \alpha \in \left( {S + {rB}}\right) \smallsetminus S \), there is a trajectory \( x \) of \( F \) on \( \lbrack 0,\infty ) \) such that \( x\left( 0\right) = \alpha \), and such that \[ {d}_{S}\left( {x\left( t\right) }\right) + {\lambda t} \leq {d}_{S}\left( \alpha \right) \] for all \( t > 0 \) up to the first \( T > 0 \) such that \( {d}_{S}\left( {x\left( T\right) }\right) = 0 \) ; i.e., such that \( x\left( T\right) \in S \) . Note that since \( \lambda > 0 \), such \( T \) must exist (uniformly for \( \alpha \) ). Once \( x\left( T\right) \in S \), then there is a trajectory beginning at \( x\left( T\right) \) which extends \( x \) and remains in \( S \) thereafter, since (1) implies that \( \left( {S, F}\right) \) is weakly invariant. 6.7. Exercise. (a) Prove that in Theorem 6.6, the parameter \( T \) can be taken to be \( r/\left( {\delta - {Kr}}\right) \) . (b) The time required to steer \( \alpha \in S + \bar{r}B \) to \( S \) is no greater than \( {d}_{S}\left( \alpha \right) /\left( {\delta - K{d}_{S}\left( \alpha \right) }\right) \), and an approach to \( S \) at rate \( \gamma \mathrel{\text{:=}} \) \( \delta - K{d}_{S}\left( \alpha \right) > 0 \) can be guaranteed, in the sense that we have \[ {d}_{S}\left( {x\left( t\right) }\right) - {d}_{S}\left( \alpha \right) \leq - {\gamma t} \] until \( S \) is reached. A sufficient condition for local attainability can also be given in tangential terms akin to those employed in proving the existence of equilibria in Theorem 4.6. 6.8. Proposition. Let \( S \) be compact, and let \( F \) be locally Lipschitz. Suppose that \[ F\left( x\right) \cap \operatorname{int}{T}_{S}^{C}\left( x\right) \neq \varnothing \forall x \in S. \] Then \( \left( {S, F}\right) \) is locally attainable. Proof. Let \( x \in S \) . Then there exists \( {v}_{x} \in F\left( x\right) \) and \( {\delta }_{x} > 0 \) such that \( {v}_{x} + {\delta }_{x}\bar{B} \subset {T}_{S}^{C}\left( x\right) \) . Given any \( \zeta \in {N}_{S}^{L}\left( x\right) \), since \[ {T}_{S}^{C}\left( x\right) = {\left( {N}_{S}^{L}\left( x\right) \right) }^{0} \] we have \[ \left\langle {{v}_{x} + {\delta }_{x}u,\zeta }\right\rangle \leq 0\;\text{ for any }u \in \bar{B} \] which implies \( \left\langle {{v}_{x},\zeta }\right\rangle \leq - {\delta }_{x}\parallel \zeta \parallel \) and consequently \( h\left( {x,\zeta }\right) \leq - {\delta }_{x}\parallel \zeta \parallel \) . This shows that the system \( \left( {S, F}\right) \) satisfies a pointwise version of condition (1), and with \( {N}_{S}^{P} \) replaced by \( {N}_{S}^{L} \) . If the system fails to satisfy (1) uniformly, then there exist sequences \( \left\{ {x}_{i}\right\} \) , \( \left\{ {\zeta }_{i}\right\} \) with \( {x}_{i} \in S,\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} = 1,{\zeta }_{i} \in {N}_{S}^{P}\left( {x}_{i}\right) \), and such that \[ h\left( {{x}_{i},{\zeta }_{i}}\right) \geq - \frac{1}{i}\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} = - \frac{1}{i}. \] By passing to subsequences, we can suppose \( {x}_{i} \rightarrow x \in S,{\zeta }_{i} \rightarrow \zeta \), where \( \zeta \in \) \( {N}_{S}^{L}\left( x\right) \) and \( \parallel \zeta \parallel = 1 \) . Since \( h\left( {x, p}\right) \) is locally Lipschitz in \( x \) (Exercise 3.3(d)) and continuous in \( p \) (as a concave real-valued function), we obtain \( h\left( {x,\zeta }\right) \geq \) 0 . But this contradicts what was proven above. Thus the system \( \left( {S, F}\right) \) satisfies (1) for some positive \( \delta \), and so is locally attainable by Theorem 6.6. ## 7 The Hamilton-Jacobi Equation and Viscosity Solutions For the first time, we now consider the issue of finding a trajectory which is best relative to a given criterion. We consider the following optimal control problem (P) \[ \text{minimize}\ell \left( {x\left( T\right) }\right) \text{subject to}\dot{x}\left( t\right) \in F\left( {x\left( t\right) }\right) \text{a.e.,}\;t \in \left\lbrack {0, T}\right\rbrack, x\left( 0\right) = {x}_{0}\text{.} \] Here \( T > 0,{x}_{0} \in {\mathbb{R}}^{n} \), and a continuous function \( \ell : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) are given. We seek therefore to minimize the endpoint cost \( \ell \left( {x\left( T\right) }\right) \) over all trajectories \( x \) of \( F \) on \( \left\lbrack {0, T}\right\rbrack \) originating at \( {x}_{0} \) . It follows from the Standing Hypotheses that problem (P) admits a solution (Exercise 3.12). We assume throughout this section that \( F \) is locally Lipschitz and autonomous. ## Verification Functions We now describe how a venerable idea in the calculus of variations, originating with Legendre, leads to sufficient conditions for optimality in control problems. Suppose that we have a feasible arc \( \bar{x} \) that we suspect of being optimal for our problem. How can we confirm that \( \bar{x} \) is a solution? Here is one way: produce a smooth \( \left( {C}^{1}\right) \) function \( \varphi \left( {t, x}\right) \) such that \[ {\varphi }_{t}\left( {t, x}\right) + \left\langle {{\varphi }_{x}\left( {t, x}\right), v}\right\rangle \geq 0\forall x,\forall t \in \left( {0, T}\right) ,\forall v \in F\left( x\right) , \] (1) \[ \varphi \left( {T, \cdot }\right) = \ell \left( \cdot \right) \] \[ \varphi \left( {0,{x}_{0}}\right) = \ell \left( {\bar{x}\left( T\right) }\right) . \] Let us see how the existence of \( \varphi \) verifies that \( \bar{x} \) is optimal. Let \( x \) be any other arc feasible for \( \left( \mathrm{P}\right) \) . Then a.e. on \( \left\lbrack {0, T}\right\rbrack \) we have \[ \frac{d}{dt}\varphi \left( {t, x\left( t\right) }\right) = {\varphi }_{t}\left( {t, x\left( t\right) }\right) + \left\langle {{\varphi }_{x}\left( {t, x\left( t\right) }\right) ,\dot{x}\left( t\right) }\right\rangle \geq 0\;\left( {\mathrm{{by}}\left( 1\right) }\right) . \] Integrating this on \( \left\lbrack {0, T}\right\rbrack \) yields \[ \varphi \left( {T, x\left( T\right) }\right) = \ell \left( {x\left( T\right) }\right) \geq \varphi \left( {0,{x}_{0}}\right) = \ell \left( {\bar{x}\left( T\right) }\right) . \] So \( \bar{x} \) gives the least possible value of \( \ell \left( {x\left( T\right) }\right) \), as required. It also follows that \( \varphi \left( {0,{x}_{0}}\right) \) is the value \( V \) of \( \left( \mathrm{P}\right) \) ; i.e., the minimum cost. In this simple argument, the Hamilton-Jacobi inequality (1) was really only used to deduce that the map \( t \rightarrow \varphi \left( {t, x\left( t\right) }\right) \) is increasing whenever \( x \) is a trajectory. In the terminology of the previous section, we want \( \left( {\varphi, F}\right) \) to be strongly increasing on \( \left( {0, T}\right) \times {\mathbb{R}}^{n} \) . Our previous results allow us to characterize this system property even when \( \varphi \) is nondifferentiable. Here is a corresponding extension of the verification argument whose proof is an adaptation of the classical one. In this and subsequent results, it is notationally convenient to use the augmented Hamiltonian \( \bar{h} \) defined by \[ \bar{h}\left( {x,\theta ,\zeta }\right) \mathrel{\text{:=}} \theta + h\left( {x,\zeta }\right) . \] 7.1. Proposition. Let \( \bar{x} \) be feasible for \( \left( \mathrm{P}\right) \), and suppose that there exists a continuous \( \varphi \left( {t, x}\right) \) on \( \left\lbrack {0, T}\right\rbrack \times {\mathbb{R}}^{n} \) satisfying \[ \bar{h}\left( {x,{\partial }_{P}\varphi \left( {t, x}\right) }\right) \geq 0\forall \left( {t, x}\right) \in \left( {0, T}\right) \times {\mathbb{R}}^{n}, \] (2) \[ \varphi \left( {T, \cdot }\right) = \ell \left( \cdot \right) \] (3) \[ \varphi \left( {0,{x}_{0}}\right) = \ell \left( {\bar{x}\left( T\right) }\right) . \] Then \( \bar{x} \) solves \( \left( \mathrm{P}\right) \), and the value of \( \left( \mathrm{P}\right) \) is \( \varphi \left( {0,{x}_{0}}\right) \) . 7.2. Exercise. Prove Proposition 7.1. This is an extension to nonsmooth \( \varphi \) of Legendre’s approach to sufficient conditions, which in the calculus of variations has also been called the "royal road of Carathéodory.” A continuous function \( \varphi \) satisfying the hypotheses (2),(3), and (4) of Proposition 7.1 is called a verification function (for \( \bar{x} \) ). The obvious question to ask at this point is how applicable the method turns out to be, or to rephrase this: Can we be sure that a verification function \( \varphi \) for \( \bar{x} \) exists when \( \bar{x} \) is optimal? And how do we find one? Considerable insight into this question arises from applying the technique of invariant embedding. Suppose that instead of the problem (P) considered above, we consider a family of problems \( P\left( {\tau ,\alpha }\right) \) parametrized by the initial data \( \left( {\tau ,\alpha }\right) \in \left\lbrack {0, T}\right\rbrack \times {\mathbb{R}}^{n} \) ; i.e., the initial condition is \[ x\left( \tau \right) = \alpha \] rather than \( x\left( 0\right) = {x}_{0} \) . Let \( V\left( {\tau ,\alpha }\right) \) denote the value of \( P\left( {\tau ,\alpha }\right) \) ; then we observe that the classical verification argument actually gives not only \( V\left( {0,{x}_{0}}\right) = \varphi \left( {0,{x}_{0}}\right) \) as noted earlier, but also \[ V\left( {\tau ,\alpha }\right) \geq \varphi \left( {\tau ,\alpha }\right) \forall \left( {\tau ,\alpha }\right) \in \lbrack 0, T) \times {\mathbb{R}}^{n}. \] We are quite naturally led to consider whether we could take \( V \) itself as the function \( \varphi \) in Proposition 7.1. We know that \( V \) is continuous on \( ( - \infty, T\rbrack \times \) \( {\mathbb{R}}^{n} \) (see Exercise 3.12). We can easily see that \( V \) indeed satisfies (2). The reason is that \( V\left( {t, x\left( t\right) }\right) \) is always increasing when \( x \) is a trajectory; the minimum of \( \ell \le
1012_(GTM178)Nonsmooth Analysis and Control Theory
80
n} \) ; i.e., the initial condition is \[ x\left( \tau \right) = \alpha \] rather than \( x\left( 0\right) = {x}_{0} \) . Let \( V\left( {\tau ,\alpha }\right) \) denote the value of \( P\left( {\tau ,\alpha }\right) \) ; then we observe that the classical verification argument actually gives not only \( V\left( {0,{x}_{0}}\right) = \varphi \left( {0,{x}_{0}}\right) \) as noted earlier, but also \[ V\left( {\tau ,\alpha }\right) \geq \varphi \left( {\tau ,\alpha }\right) \forall \left( {\tau ,\alpha }\right) \in \lbrack 0, T) \times {\mathbb{R}}^{n}. \] We are quite naturally led to consider whether we could take \( V \) itself as the function \( \varphi \) in Proposition 7.1. We know that \( V \) is continuous on \( ( - \infty, T\rbrack \times \) \( {\mathbb{R}}^{n} \) (see Exercise 3.12). We can easily see that \( V \) indeed satisfies (2). The reason is that \( V\left( {t, x\left( t\right) }\right) \) is always increasing when \( x \) is a trajectory; the minimum of \( \ell \left( {x\left( T\right) }\right) \) can only be greater or equal starting from an intermediate point \( \left( {{t}^{\prime }, x\left( {t}^{\prime }\right) }\right) \) than it was from an earlier "less committed" point \( \left( {\tau, x\left( \tau \right) }\right) \) . (This is an instance of the logic known as the principle of optimality.) This strong increase property of the system \( \left( {V, F}\right) \), by Proposition 6.5, implies the Hamilton-Jacobi inequality (2) on all \( \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \) . Finally, it is clear that \( V \) satisfies (3), and satisfies (4) iff \( \bar{x} \) is optimal. We obtain therefore the following satisfying justification of the verification method: 7.3. Proposition. A feasible arc \( \bar{x} \) is optimal iff there exists a continuous verification function for \( \bar{x} \) ; the value function \( V \) is one such verification function for any optimal arc. 7.4. Exercise. We set \( n = 1, F\left( x\right) = \left\lbrack {-\left| x\right| ,\left| x\right| }\right\rbrack ,\ell \left( x\right) = x, T = 1 \) . (a) Calculate \( V\left( {\tau ,\alpha }\right) \) for any \( \tau \leq 1 \), and verify that it satisfies the properties of a verification function. Note that \( V \) is nondifferentiable. (b) Find a different verification function confirming the optimality of the arc that solves \( P\left( {0,0}\right) \) . There are in general many possible verification functions for a given optimal \( \bar{x} \) . We have seen in this section, however, how naturally the value function \( V \) arises in connection with the verification method, and hence the associated Hamilton-Jacobi inequality. Might it be possible to establish an even closer relationship, perhaps even a characterization of \( V \), in Hamilton–Jacobi terms? ## The Proximal Hamilton-Jacobi Equation The following theorem shows that the value function is the unique continuous solution of a suitable generalization of the classical Hamilton-Jacobi, whose general form is \[ {\varphi }_{t} + H\left( {x,{\varphi }_{x}}\right) = 0 \] with boundary condition. Recall that \( \bar{h}\left( {x,\theta ,\zeta }\right) \) is defined as \( \theta + h\left( {x,\zeta }\right) \) . We call a function \( \varphi \) that satisfies (5) below a proximal solution of the Hamilton-Jacobi equation (for \( h \) ). 7.5. Theorem. There is a unique continuous function \( \varphi : ( - \infty, T\rbrack \times {\mathbb{R}}^{n} \rightarrow \) \( \mathbb{R} \) satisfying \[ \bar{h}\left( {x,{\partial }_{P}\varphi \left( {t, x}\right) }\right) = 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n}, \] (4) \[ \ell \left( x\right) = \varphi \left( {T, x}\right) \forall x \in {\mathbb{R}}^{n}. \] (5) That function is the value function \( V \) . Proof. That \( V \) satisfies (6) was noted earlier, as well as "half" of (5); there remains to show \[ \bar{h}\left( {x,{\partial }_{P}V\left( {t, x}\right) }\right) \leq 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n}. \] (6) But whenever \( V\left( {\tau ,\alpha }\right) \) is finite, there is an optimal arc \( \bar{x} \) for the problem \( P\left( {\tau ,\alpha }\right) \), and along \( \bar{x}, V \) is constant (i.e., \( t \rightarrow V\left( {t,\bar{x}\left( t\right) }\right) \) is constant on \( \left\lbrack {\tau, T}\right\rbrack \) ). Thus the system \( \left( {V, F}\right) \) is weakly decreasing relative to \( t \in \left( {-\infty, T}\right) \), so that (7) holds by Exercise 6.4. We have shown that \( V \) satisfies (5) and (6). Now let \( \varphi \) be any other function as described in the theorem. Let us show first that \( V \leq \varphi \) . To this end, let \( \left( {\tau ,\alpha }\right) \) be any point with \( \tau < T \) . Then the system \( \left( {\varphi, F}\right) \) is weakly decreasing relative to \( t < T \), so that there is a trajectory \( x \) on \( \left\lbrack {\tau, T}\right\rbrack \) with \( x\left( \tau \right) = \alpha \) such that \[ \varphi \left( {t, x\left( t\right) }\right) \leq \varphi \left( {\tau ,\alpha }\right) \forall t \in \lbrack \tau, T). \] Letting \( t \uparrow T \), we derive \( \ell \left( {x\left( T\right) }\right) = \varphi \left( {T, x\left( T\right) }\right) \leq \varphi \left( {\tau ,\alpha }\right) \), which implies \( V\left( {\tau ,\alpha }\right) \leq \varphi \left( {\tau ,\alpha }\right) . \) We now proceed to show \( V \geq \varphi \) . Let \( \left( {\tau ,\alpha }\right) \) be any point with \( \tau < T \) . Then there exists a trajectory \( \bar{x} \) optimal for \( P\left( {\tau ,\alpha }\right) \) . Because \( \left( {\varphi, F}\right) \) is strongly increasing, we derive \[ \varphi \left( {T,\bar{x}\left( T\right) }\right) \geq \varphi \left( {\tau ,\alpha }\right) \] But \( \varphi \left( {T,\bar{x}\left( T\right) }\right) = \ell \left( {\bar{x}\left( T\right) }\right) = V\left( {\tau ,\alpha }\right) \), which completes the proof. The proof actually establishes two comparison theorems that we proceed to note formally: 7.6. Corollary. Let \( \varphi : ( - \infty, T\rbrack \times {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be continuous and satisfy: (a) \( \bar{h}\left( {x,{\partial }_{P}\varphi \left( {t, x}\right) }\right) \leq 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \) ; and (b) \( \ell \left( x\right) \leq \varphi \left( {T, x}\right) \forall x \in {\mathbb{R}}^{n} \) . Then \( \varphi \geq V \) . 7.7. Corollary. Let \( \varphi : ( - \infty, T\rbrack \times {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be continuous and satisfy: (a) \( \bar{h}\left( {x,{\partial }_{P}\varphi \left( {t, x}\right) }\right) \geq 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \) ; and (b) \( \ell \left( x\right) \geq \varphi \left( {T, x}\right) \forall x \in {\mathbb{R}}^{n} \) . Then \( \varphi \leq V \) . We remark that Corollary 7.6 is valid without the Lipschitz hypothesis on \( F \), in contrast to its companion. ## Minimax Solutions It is possible to express the extended Hamilton-Jacobi equation in terms of other constructs of nonsmooth analysis besides proximal subgradients, for example, via subderivates. Subbotin has called the concept of solution that appears below a minimax solution. 7.8. Proposition. \( V \) is the unique continuous function \( \varphi : ( - \infty, T\rbrack \times {\mathbb{R}}^{n} \rightarrow \) \( \mathbb{R} \) satisfying: (a) \( \mathop{\inf }\limits_{{v \in F\left( x\right) }}{D\varphi }\left( {t, x;1, v}\right) \leq 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \) ; (b) \( \mathop{\sup }\limits_{{v \in F\left( x\right) }}{D\varphi }\left( {t, x; - 1, - v}\right) \leq 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \) ; and (c) \( \varphi \left( {T, \cdot }\right) = \ell \left( \cdot \right) \) . Proof. It suffices to prove that the two conditions (a) of Corollary 7.6 and of Proposition 7.8 are equivalent, and that conditions (a) of Corollary 7.7 and (b) of the proposition are equivalent. First, let (a) of Proposition 7.8 hold, and let \( \left( {\theta ,\zeta }\right) \) belong to \( {\partial }_{P}\varphi \left( {t, x}\right) \) . Let \( \varepsilon > 0 \) be given, and let \( v \) in \( F\left( x\right) \) satisfy \( {D\varphi }\left( {t, x;1, v}\right) < \varepsilon \) . Then \[ \varepsilon > {D\varphi }\left( {t, x;1, v}\right) \geq \langle \left( {\theta ,\zeta }\right) ,\left( {1, v}\right) \rangle \geq \bar{h}\left( {x,\theta ,\zeta }\right) . \] We deduce (a) of Corollary 7.6, since \( \varepsilon \) is arbitrary. Now let (a) of Corollary 7.6 hold, and suppose that for some \( \varepsilon > 0 \), we had \[ {D\varphi }\left( {t, x;1, v}\right) > \varepsilon \forall v \in F\left( x\right) . \] Since \( {D\varphi } \) is lower semicontinuous in \( v \) (Exercise 3.4.1(e)), for some \( \delta > 0 \) we have \[ {D\varphi }\left( {t, x;1, v}\right) > \frac{\varepsilon }{2}\forall v \in F\left( x\right) + {\delta B}. \] By Subbotin’s Theorem 3.4.2 this implies the existence, for any \( r > 0 \), of \( \left( {\theta ,\zeta }\right) \in {\partial }_{P}\varphi \left( {{t}^{\prime },{x}^{\prime }}\right) \), where \( \left( {{t}^{\prime },{x}^{\prime }}\right) \) is of distance less than \( r \) from \( \left( {t, x}\right) \), such that \[ \langle \left( {\theta ,\zeta }\right) ,\left( {1, v}\right) \rangle > \frac{\varepsilon }{3}\forall v \in F\left( x\right) + {\delta B}. \] Thus, as soon as \( F\left( {x}^{\prime }\right) \) lies within \( F\left( x\right) + {\delta B} \) we deduce \[ \bar{h}\left( {{x}^{\prime },\theta ,\zeta }\right) \geq \frac{\varepsilon }{3}\text{ for some }\left( {\theta ,\zeta }\right) \in {\partial }_{P}\varphi \left( {{t}^{\prime },{x}^{\prime }}\right) , \] contradicting (a) of Corollary 7.6 as required. Thi
1012_(GTM178)Nonsmooth Analysis and Control Theory
81
cise 3.4.1(e)), for some \( \delta > 0 \) we have \[ {D\varphi }\left( {t, x;1, v}\right) > \frac{\varepsilon }{2}\forall v \in F\left( x\right) + {\delta B}. \] By Subbotin’s Theorem 3.4.2 this implies the existence, for any \( r > 0 \), of \( \left( {\theta ,\zeta }\right) \in {\partial }_{P}\varphi \left( {{t}^{\prime },{x}^{\prime }}\right) \), where \( \left( {{t}^{\prime },{x}^{\prime }}\right) \) is of distance less than \( r \) from \( \left( {t, x}\right) \), such that \[ \langle \left( {\theta ,\zeta }\right) ,\left( {1, v}\right) \rangle > \frac{\varepsilon }{3}\forall v \in F\left( x\right) + {\delta B}. \] Thus, as soon as \( F\left( {x}^{\prime }\right) \) lies within \( F\left( x\right) + {\delta B} \) we deduce \[ \bar{h}\left( {{x}^{\prime },\theta ,\zeta }\right) \geq \frac{\varepsilon }{3}\text{ for some }\left( {\theta ,\zeta }\right) \in {\partial }_{P}\varphi \left( {{t}^{\prime },{x}^{\prime }}\right) , \] contradicting (a) of Corollary 7.6 as required. This completes the proof that the corresponding (a) parts are equivalent. The remainder of the proof is similar to the above, and is left as an exercise. 7.9. Exercise. Prove that condition (b) of Proposition 7.8 and condition (a) of Corollary 7.7 are equivalent. ## Viscosity Solutions We now establish that the value function is also the unique viscosity solution of the Hamilton-Jacobi boundary-value problem. This celebrated solution concept, developed by M. Crandall and P. L. Lions, is bilateral like minimax solutions, but even more so, in that it uses both subdifferen-tials and superdifferentials. Of course, given the uniqueness of the solution, it is evident that the proximal, minimax, and viscosity solution concepts all coincide in our present setting. Let us recall the \( D \) -subdifferential \( {\partial }_{D}f \) of \( §{3.4} \), and the characterization provided by Proposition 3.4.12: \( \zeta \) belongs to \( {\partial }_{D}f\left( x\right) \) iff there is a function \( g \) differentiable at \( x \), with \( {g}^{\prime }\left( x\right) = \zeta \), such that \( f - g \) has a local minimum at \( x \) . The \( D \) -superdifferential \( {\partial }^{D}f\left( x\right) \) is defined analogously, with \( f - g \) having a local maximum at \( x \) . 7.10. Proposition. \( V \) is the unique continuous function \( \varphi : ( - \infty, T\rbrack \times \) \( {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) satisfying: (a) \( \bar{h}\left( {x,{\partial }_{D}\varphi \left( {t, x}\right) }\right) \leq 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \) ; (b) \( \bar{h}\left( {x,{\partial }^{D}\varphi \left( {t, x}\right) }\right) \geq 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \) ; and (c) \( \varphi \left( {T, \cdot }\right) = \ell \left( \cdot \right) \) . Proof. It is an immediate consequence of Proposition 3.4.5 that (a) of this proposition and (a) of Corollary 7.6 are equivalent. It suffices then to prove the equivalence of (b) of the proposition to (a) of Corollary 7.7. Now the latter condition is equivalent to the strong increase of the system \( \left( {\varphi, F}\right) \) on \( \left( {-\infty, T}\right) \), as we know. This in turn is equivalent to the strong decrease of the system \( \left( {-\varphi, F}\right) \) on \( \left( {-\infty, T}\right) \), which in turn is equivalent to \[ \theta + H\left( {x,\theta ,\zeta }\right) \leq 0\forall \left( {\theta ,\zeta }\right) \in {\partial }_{P}\left( {-\varphi }\right) \left( {t, x}\right) ,\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n}, \] by Exercise 6.4. Invoking Proposition 3.4.5 again, this is equivalent to \[ \theta + H\left( {x,\theta ,\zeta }\right) \leq 0\forall \left( {\theta ,\zeta }\right) \in {\partial }_{D}\left( {-\varphi }\right) \left( {t, x}\right) ,\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n}. \] Since \( {\partial }^{D}\varphi \) is precisely \( - {\partial }_{D}\left( {-\varphi }\right) \), this last condition coincides with (b) of Proposition 7.10, and we are done. We remark that in terms of the usual conventions in the literature of viscosity solutions, Proposition 7.10 asserts that \( V \) is the viscosity solution of the equation \( - \bar{h}\left( {x,\nabla \varphi }\right) = 0 \) rather than \( \bar{h}\left( {x,\nabla \varphi }\right) = 0 \) . The minus sign here makes a difference, as it clearly must in any bilateral mode of definition. In the unilateral proximal setting, the difference resides in the fact that it is \( {\partial }_{P}\varphi \) figuring in (5), and not \( {\partial }^{P}\varphi \) . ## 7.11. Exercise. (a) Let \( \varphi \) be a continuous function satisfying (5). Prove that at any point \( \left( {t, x}\right) \) at which \( \varphi \) is differentiable, we have \[ {\varphi }_{t}\left( {t, x}\right) + h\left( {x,{\varphi }_{x}\left( {t, x}\right) }\right) = 0. \] (7) Deduce that if \( \varphi \) is locally Lipschitz, then this holds a.e. on \( \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \) . This defines an earlier notion of generalized solution to the Hamilton-Jacobi equation, one that could be called an almost everywhere solution. We proceed to show that almost everywhere solutions are not necessarily unique, and hence differ from the proximal ones. (b) We set \( n = 1, F\left( x\right) = \left\lbrack {-1,1}\right\rbrack ,\ell \left( x\right) = \left| x\right| \), and \( T = 1 \) . Show that the value function \( V \) is given by \[ V\left( {t, x}\right) = \max \{ \left| x\right| + t - 1,0\} . \] (c) Confirm that \( V \) is a Lipschitz almost everywhere solution of the Hamilton-Jacobi equation with boundary condition \( V\left( {1, x}\right) = \) \( \left| x\right| \) . Show that another is \( \varphi \left( {t, x}\right) \mathrel{\text{:=}} \left| x\right| + t - 1 \), but that \( \varphi \) fails to satisfy (5), in contrast to \( V \) . (d) Show that both \( \varphi \) and \( V \) satisfy (5) if \( {\partial }_{P} \) is replaced by \( {\partial }^{P} \) . (e) If \( \varphi \) is differentiable at every point and satisfies (8) on \( \left( {-\infty, T}\right) \times \) \( {\mathbb{R}}^{n} \) without exception, together with the boundary condition, prove that \( \varphi = V \) on \( ( - \infty, T\rbrack \times {\mathbb{R}}^{n} \) . ## 8 Feedback Synthesis from Semisolutions The Hamilton-Jacobi inequality \[ \bar{h}\left( {x,{\partial }_{P}\varphi \left( {t, x}\right) }\right) \leq 0 \] together with the boundary condition \[ \varphi \left( {T, \cdot }\right) \geq \ell \left( \cdot \right) \] defines what is called a semisolution. Such a function can be useful in producing upper bounds for the value \( V \) of our optimal control problem; as we saw, these conditions imply \( V \leq \varphi \) (Corollary 7.6). Thus for each \( \left( {\tau ,\alpha }\right) \in ( - \infty, T\rbrack \times {\mathbb{R}}^{n} \) there is a trajectory \( \bar{x} \) with \( x\left( \tau \right) = \alpha \) and \( \ell \left( {\bar{x}\left( T\right) }\right) \leq \) \( \varphi \left( {\tau ,\alpha }\right) \) . We address now the issue of actually constructing such a trajectory. In the special case in which \( \varphi = V \), this becomes the problem of finding optimal trajectories. It is interesting to recall the classical approach to this issue. Assuming that \( \varphi \) is smooth, this would direct us to select for each \( \left( {t, x}\right) \) a point \( \bar{v}\left( {t, x}\right) \) in \( F\left( x\right) \) at which the minimum defining \( \bar{h}\left( {x,\nabla \varphi \left( {t, x}\right) }\right) \) is attained; i.e., such that \[ {\varphi }_{t}\left( {t, x}\right) + \left\langle {{\varphi }_{x}\left( {t, x}\right) ,\bar{v}}\right\rangle = \bar{h}\left( {x,\nabla \varphi \left( {t, x}\right) }\right) . \] Then, we proceed to define a trajectory \( \bar{x} \) via \[ \dot{\bar{x}}\left( t\right) = \bar{v}\left( {t,\bar{x}\left( t\right) }\right) ,\;\bar{x}\left( \tau \right) = \alpha . \] If all this is possible, we derive \[ \ell \left( {\bar{x}\left( T\right) }\right) \leq \varphi \left( {\tau ,\alpha }\right) \] as follows: \[ \ell \left( {\bar{x}\left( T\right) }\right) - \varphi \left( {\tau ,\alpha }\right) \leq \varphi \left( {T,\bar{x}\left( T\right) }\right) - \varphi \left( {\tau ,\alpha }\right) \] \[ = {\int }_{0}^{T}\frac{d}{dt}\varphi \left( {t,\bar{x}\left( t\right) }\right) {dt} \] \[ = {\int }_{0}^{T}\left\{ {{\varphi }_{t}\left( {t,\bar{x}\left( t\right) }\right) + \left\langle {{\varphi }_{x}\left( {t,\bar{x}\left( t\right) }\right) ,\dot{\bar{x}}\left( t\right) }\right\rangle }\right\} {dt} \] \[ = {\int }_{0}^{T}\bar{h}\left( {\bar{x}\left( t\right) ,\nabla \varphi \left( {t,\bar{x}\left( t\right) }\right) }\right) {dt} \leq 0. \] The difficulties with this "dynamic programming" approach are intrinsic (smoothness of \( \varphi \), regularity of \( \bar{v} \), existence of \( \bar{x} \) ), but it is of note that it attempts to construct a feedback giving rise to the required trajectory. Proximal aiming allows us to rescue the approach in essentially these terms, and for merely lower semicontinuous semisolutions. The integration step above is still not possible, but proximal methods produce the required system monotonicity. The following result is obtained. It is not just an abstract existence theorem: we construct a feedback \( \bar{v} \) quite explicitly. 8.1. Theorem. Let \( F \) be locally Lipschitz, and let \( \varphi \in \mathcal{F}\left( {\left( {-\infty, T}\right) \times {\mathbb{R}}^{n}}\right) \) satisfy \[ \bar{h}\left( {x,{\partial }_{P}\varphi \left( {t, x}\right) }\right) \leq 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \] and \[ \ell \left( x\right) \leq \mathop{\liminf }\limits_{\substack{{{t}^{\prime } \uparrow T} \\ {{x}^{\prime } \rightarrow x} }}\varphi \left( {{
1012_(GTM178)Nonsmooth Analysis and Control Theory
82
\( \bar{v} \), existence of \( \bar{x} \) ), but it is of note that it attempts to construct a feedback giving rise to the required trajectory. Proximal aiming allows us to rescue the approach in essentially these terms, and for merely lower semicontinuous semisolutions. The integration step above is still not possible, but proximal methods produce the required system monotonicity. The following result is obtained. It is not just an abstract existence theorem: we construct a feedback \( \bar{v} \) quite explicitly. 8.1. Theorem. Let \( F \) be locally Lipschitz, and let \( \varphi \in \mathcal{F}\left( {\left( {-\infty, T}\right) \times {\mathbb{R}}^{n}}\right) \) satisfy \[ \bar{h}\left( {x,{\partial }_{P}\varphi \left( {t, x}\right) }\right) \leq 0\forall \left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \] and \[ \ell \left( x\right) \leq \mathop{\liminf }\limits_{\substack{{{t}^{\prime } \uparrow T} \\ {{x}^{\prime } \rightarrow x} }}\varphi \left( {{t}^{\prime },{x}^{\prime }}\right) \forall x \in {\mathbb{R}}^{n}. \] Then for given \( \left( {\tau ,\alpha }\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} \), there exists a feedback selection \( \bar{v} \) of \( F \) with the property that every Euler solution \( \bar{x} \) of the initial-value problem \[ \dot{x} = \bar{v}\left( {t, x}\right) ,\;x\left( \tau \right) = \alpha , \] satisfies \( \ell \left( {\bar{x}\left( T\right) }\right) \leq \varphi \left( {\tau ,\alpha }\right) \) . Proof. Let us consider the set \[ S \mathrel{\text{:=}} \left\{ {\left( {t, x}\right) \in \left( {-\infty, T}\right) \times {\mathbb{R}}^{n} : \varphi \left( {t, x}\right) \leq \varphi \left( {\tau ,\alpha }\right) }\right\} \cup \left\{ {\left( {t, x}\right) : t \geq T, x \in {\mathbb{R}}^{n}}\right\} , \] and note that \( S \) is closed, since \( \varphi \) is lower semicontinuous, and also that the system \( \left( {S,\bar{F}}\right) \) is weakly invariant, where \( \bar{F}\left( x\right) \mathrel{\text{:=}} \{ 1\} \times F\left( x\right) \) . This is essentially a restatement of the fact that \( \left( {\varphi, F}\right) \) is weakly decreasing on \( t < T \), which follows from Exercise 6.4. We now take the feedback selection \( {g}_{P} \) for \( \bar{F} \) defined in Theorem 3.4; necessarily, it is of the form \( \left( {1,\bar{v}\left( {t, x}\right) }\right) \), where \( \bar{v} \) is a feedback selection for \( F \) . It follows that any Euler solution of \( \dot{x} = \bar{v}\left( {t, x}\right), x\left( \tau \right) = \alpha \) is such that \( \left( {t, x\left( t\right) }\right) \in S\forall t \geq \tau \) . Thus \[ \varphi \left( {\tau ,\alpha }\right) \geq \mathop{\liminf }\limits_{{\varepsilon \downarrow 0}}\varphi \left( {T - \varepsilon, x\left( {T - \varepsilon }\right) }\right) \geq \ell \left( {x\left( T\right) }\right) . \] We recall that (as shown in the proof of Theorem 3.4), the feedback \( \bar{v}\left( {t, x}\right) \) is constructed as follows: for given \( \left( {t, x}\right) \), we first find any point \( \left( {{t}^{\prime },{x}^{\prime }}\right) \in \) \( {\operatorname{proj}}_{S}\left( {t, x}\right) \) ; next, we locate \( v \in F\left( {x}^{\prime }\right) \) minimizing the function \( v \mapsto \langle v, x - \) \( {x}^{\prime }\rangle \) over \( F\left( {x}^{\prime }\right) \) . Finally, we take \( \bar{v}\left( {t, x}\right) \) in \( F\left( x\right) \) closest to \( v \) . This same \( \bar{v} \) will provide trajectories \( x \) satisfying \( \ell \left( {x\left( T\right) }\right) \leq \varphi \left( {\tau ,\alpha }\right) \) from any initial data \( \left( {{\tau }^{\prime },{\alpha }^{\prime }}\right) \) for which \( \varphi \left( {{\tau }^{\prime },{\alpha }^{\prime }}\right) \leq \varphi \left( {\tau ,\alpha }\right) \) . In general \( \bar{v} \) is discontinuous, and \( \bar{v}\left( {t, x}\right) \) is an arbitrary element of \( F\left( {t, x}\right) \) when \( {x}^{\prime } = x \) . We remark that a refinement of this construction can be made so as to define, for any compact subset \( C \) and \( \varepsilon > 0 \), a feedback selection generating trajectories \( x \) that satisfy \( \ell \left( {x\left( T\right) }\right) \leq \varphi \left( {\tau ,\alpha }\right) + \varepsilon \) whenever \( \left( {\tau ,\alpha }\right) \) lies in \( C \), though we shall omit this topic. 8.2. Exercise. For \( \varphi = V \) in Exercise \( {7.11}\left( \mathrm{\;b}\right) \), and for \( \left( {\tau ,\alpha }\right) = \) \( \left( {0,0}\right) \), describe the set \( S \) and the optimal feedback \( \bar{v} \) . Sketch typical directions of \( \bar{v} \) in the \( \left( {t, x}\right) \) plane. Note that \( \bar{v} \) cannot be defined so as to be continuous on \( ( - \infty ,1\rbrack \times \mathbb{R} \) . Prove that no continuous feedback can be optimal. 8.3. Exercise. With \( n = 1 \), set \( h\left( {x, p}\right) \mathrel{\text{:=}} - \left| {xp}\right| \), and let \[ \varphi \left( {t, x}\right) = \left\{ \begin{array}{ll} x{e}^{t - 1} & \text{ if }x \geq 0 \\ x{e}^{1 - t} & \text{ if }x \leq 0 \end{array}\right. \] Prove that \( \varphi \) is the unique proximal/minimax/viscosity solution on \( ( - \infty ,1\rbrack \) of the Hamilton-Jacobi boundary-value problem \[ {\varphi }_{t} + h\left( {x,{\varphi }_{x}}\right) = 0,\;\varphi \left( {1, y}\right) = y\forall y. \] (Which of the three solution concepts is easiest to use in order to do this?) Why does this immediately imply that \( \varphi = V \), where \( V \) is the value function of Exercise 7.4? ## 9 Necessary Conditions for Optimal Control We have studied in \( §7 \) a verification method that can confirm the optimality of a suspect, and we have shown how to calculate optimal arcs by feedback in \( §8 \), if the value function is known (or suboptimal arcs, if a semisolution is at hand). But we still lack the necessary conditions that can be used to identify potential optimal solutions. That is the subject of this section, in which is considered the following optimal control problem (P): \[ \text{minimize}\left\{ {\ell \left( {x\left( b\right) }\right) : \dot{x} \in F\left( x\right) \text{a.e.,}x\left( a\right) = {x}_{0}}\right\} \text{.} \] That is, we wish as before to minimize \( \ell \left( {x\left( b\right) }\right) \) over the trajectories \( x \) on \( \left\lbrack {a, b}\right\rbrack \) with initial-value \( {x}_{0} \) . The interval \( \left\lbrack {a, b}\right\rbrack \) is given, together with \( {x}_{0} \in {\mathbb{R}}^{n} \) . We assume that \( \ell \) is a locally Lipschitz function, and that the autonomous multifunction \( F \) is locally Lipschitz. We will say that \( x \) is a local solution to \( \left( \mathrm{P}\right) \) if, for some \( {\varepsilon }_{0} > 0 \), we have \[ \ell \left( {x\left( b\right) }\right) \leq \ell \left( {y\left( b\right) }\right) \] whenever \( y \) is a trajectory on \( \left\lbrack {a, b}\right\rbrack \) satisfying \( y\left( a\right) = {x}_{0} \) as well as \[ \parallel y - x{\parallel }_{\infty } \leq {\varepsilon }_{0} \] Recall that the upper Hamiltonian \( H \) corresponding to \( F \) is the function \( H \) given by \[ H\left( {x, p}\right) \mathrel{\text{:=}} \max \{ \langle p, v\rangle : v \in F\left( x\right) \} . \] In the following set of necessary conditions for a local minimum, the first conclusion is known as the Hamiltonian inclusion; the condition at (b) is called the transversality condition. 9.1. Theorem. Let \( x \) be a local solution to the optimal control problem (P). Then there exists an arc \( p \) on \( \left\lbrack {a, b}\right\rbrack \) which, together with \( x \), satisfies \[ \left( {-\dot{p}\left( t\right) ,\dot{x}\left( t\right) }\right) \in {\partial }_{C}H\left( {x\left( t\right), p\left( t\right) }\right) \text{ a.e.,}\;a \leq t \leq b, \] \[ - p\left( b\right) \in {\partial }_{L}\ell \left( {x\left( b\right) }\right) \text{.} \] Proof. This is a long and fairly involved proof, in which nonsmooth calculus plays the major role. We will require the following facts regarding \( {\partial }_{C}H \) ; the Generalized Gradient Formula 2.8.1 is helpful in proving them: 9.2. Exercise. \( H \) is locally Lipschitz, and if \( \left( {q, v}\right) \) belongs to \( {\partial }_{C}H\left( {x, p}\right) \) then: (a) \( v \in F\left( x\right) \) and \( \langle p, v\rangle = H\left( {x, p}\right) \) . (b) \( \parallel q\parallel \leq K\parallel p\parallel \), where \( K \) is a local Lipschitz constant for \( F \) . (c) For any \( \lambda \geq 0,\left( {{\lambda q}, v}\right) \in {\partial }_{C}H\left( {x,{\lambda p}}\right) \) . (d) For any \( w \in F\left( x\right) \), we have \( \left( {0, w}\right) \in {\partial }_{C}H\left( {x,0}\right) \) . There is no loss of generality, we claim, in assuming that \( F \) is globally Lipschitz and bounded. For there is an a priori bound \( M \) on \( \parallel x{\parallel }_{\infty } \) for any trajectory \( x \) on \( \left\lbrack {a, b}\right\rbrack \) with \( x\left( a\right) = {x}_{0} \) (see Exercise 1.5); let us redefine \( F\left( x\right) \) for \( \parallel x\parallel > {2M} \) (only) as follows: \[ \widetilde{F}\left( x\right) = F\left( \frac{2Mx}{\parallel x\parallel }\right) . \] Then \( \widetilde{F} \) satisfies the Standing Hypotheses, and can be shown to be globally Lipschitz and bounded. Furthermore, the arc that solved the initial problem continues to solve the problem in which \( \widetilde{F} \) replaces \( F \), since the trajectories from \( {x}_{0} \) are the same. Finally, observe that \( H \) and \( \widetilde{H} \) coincide in a neighborhood of the optimal arc, and therefore so do \( {\partial }_{C}H \) and \( {\partial }_{C}\widetilde{H} \) . Thus the conclusion of the theorem is indifferent to whether \( F \) or \( \widetilde{F} \) is employed. To summarize, we assume that for some \( K > 0, F\left( x\right) \subset {KB} \) for all \( x \), and \( F \) is global
1012_(GTM178)Nonsmooth Analysis and Control Theory
83
) on \( \left\lbrack {a, b}\right\rbrack \) with \( x\left( a\right) = {x}_{0} \) (see Exercise 1.5); let us redefine \( F\left( x\right) \) for \( \parallel x\parallel > {2M} \) (only) as follows: \[ \widetilde{F}\left( x\right) = F\left( \frac{2Mx}{\parallel x\parallel }\right) . \] Then \( \widetilde{F} \) satisfies the Standing Hypotheses, and can be shown to be globally Lipschitz and bounded. Furthermore, the arc that solved the initial problem continues to solve the problem in which \( \widetilde{F} \) replaces \( F \), since the trajectories from \( {x}_{0} \) are the same. Finally, observe that \( H \) and \( \widetilde{H} \) coincide in a neighborhood of the optimal arc, and therefore so do \( {\partial }_{C}H \) and \( {\partial }_{C}\widetilde{H} \) . Thus the conclusion of the theorem is indifferent to whether \( F \) or \( \widetilde{F} \) is employed. To summarize, we assume that for some \( K > 0, F\left( x\right) \subset {KB} \) for all \( x \), and \( F \) is globally Lipschitz of rank \( K \) . We will simplify notation by taking \( \left\lbrack {a, b}\right\rbrack = \left\lbrack {0,1}\right\rbrack \), and \( {x}_{0} = 0 \), and we denote the optimal trajectory by \( \bar{x} \), and its derivative by \( \bar{v} \) . Let us now proceed to define two subsets of \( X \mathrel{\text{:=}} {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \) : \[ S \mathrel{\text{:=}} \left\{ {v \in X : v\left( t\right) \in F\left( {{\int }_{0}^{t}v\left( s\right) {ds}}\right) \text{ a.e. }}\right\} , \] \[ \sum \mathrel{\text{:=}} \left\{ {v \in X : {\begin{Vmatrix}\bar{x} - {\int }_{0}^{t}v\left( s\right) ds\end{Vmatrix}}_{\infty } \leq {\varepsilon }_{0}}\right\} . \] We denote by \( \widetilde{\ell } \) the following function on \( X \) : \[ \widetilde{\ell }\left( v\right) \mathrel{\text{:=}} \ell \left( {{\int }_{0}^{1}v\left( s\right) {ds}}\right) . \] We ask the reader to verify that \( \bar{v} \) minimizes \( \widetilde{\ell }\left( v\right) \) over \( v \in S \cap \sum \), and to provide the following facts: 9.3. Exercise. (a) \( S \) is a closed subset of \( X \) . (b) \( \bar{v} \in \operatorname{int}\sum \) . (c) \( \widetilde{\ell } \) is locally Lipschitz. It follows from the above that we have \[ 0 \in {\partial }_{P}\left\{ {\widetilde{\ell } + {I}_{S}}\right\} \left( \bar{v}\right) \] and from the Proximal Sum Rule 1.8.2 that, for any \( \varepsilon > 0 \), for some \( {v}_{1} \) and \( {v}_{2} \) within \( \varepsilon \) of \( \bar{v} \), we have \[ 0 \in {\partial }_{P}\widetilde{\ell }\left( {v}_{1}\right) + {N}_{S}^{P}\left( {v}_{2}\right) + {\varepsilon B}. \] (1) The bulk of our effort will go toward the calculation of \( {N}_{S}^{P} \) . To that end, it is useful to introduce the following subset \( C \) of \( X \times X \) : \[ C \mathrel{\text{:=}} \{ \left( {u, v}\right) \in X \times X : \left( {u\left( t\right), v\left( t\right) }\right) \in \operatorname{graph}\left( F\right) \text{ a.e. }\} . \] Before going on to the study of \( {N}_{S}^{P} \), we will need: 9.4. Lemma. Let \( \left( {\zeta ,\xi }\right) \in {N}_{C}^{P}\left( {u, v}\right) \) . Then \[ \left( {-\zeta \left( t\right), v\left( t\right) }\right) \in {\partial }_{C}H\left( {u\left( t\right) ,\xi \left( t\right) }\right) \text{ a.e.,}\;0 \leq t \leq 1. \] Proof. To begin, observe that by Proposition 3.5.7, there exists \( \sigma > 0 \) such that for almost all \( t \in \left\lbrack {0, T}\right\rbrack \), we have \[ \left\langle {\zeta \left( t\right) ,{u}^{\prime } - u\left( t\right) }\right\rangle + \left\langle {\xi \left( t\right) ,{v}^{\prime } - v\left( t\right) }\right\rangle \] \[ \leq \sigma {\begin{Vmatrix}\left( {u}^{\prime } - u\left( t\right) ,{v}^{\prime } - v\left( t\right) \right. \end{Vmatrix}}^{2}\forall \left( {{u}^{\prime },{v}^{\prime }}\right) \in \operatorname{graph}\left( F\right) . \] (2) Let us fix a value of \( t \) for which (2) holds. By setting \( {u}^{\prime } = u\left( t\right) \) in (2), we derive that for any \( {v}^{\prime } \in F\left( {u\left( t\right) }\right) \) , \[ \left\langle {\xi \left( t\right) ,{v}^{\prime }}\right\rangle \leq \langle \xi \left( t\right), v\left( t\right) \rangle + \sigma {\begin{Vmatrix}{v}^{\prime } - v\left( t\right) \end{Vmatrix}}^{2}. \] In other terms, \( \xi \left( t\right) \) lies in \( {N}_{F\left( {u\left( t\right) }\right) }^{P}\left( {v\left( t\right) }\right) \) . Since \( F\left( {u\left( t\right) }\right) \) is a convex set, this is equivalent to \[ \left\langle {\xi \left( t\right) ,{v}^{\prime }}\right\rangle \leq \langle \xi \left( t\right), v\left( t\right) \rangle \forall {v}^{\prime } \in F\left( {u\left( t\right) }\right) . \] It follows that we have \[ H\left( {u\left( t\right) ,\xi \left( t\right) }\right) = \langle \xi \left( t\right), v\left( t\right) \rangle . \] (3) We proceed to define a useful function \( g : {\mathbb{R}}^{n} \times {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) as follows: \[ g\left( {x, p}\right) \mathrel{\text{:=}} \langle \xi \left( t\right) - p, v\left( t\right) \rangle + \frac{\parallel (\xi \left( t\right) - p{\parallel }^{2}}{4\sigma } \] \[ + \langle \zeta \left( t\right), x - u\left( t\right) \rangle - \sigma \parallel x - u\left( t\right) {\parallel }^{2} + H\left( {x, p}\right) . \] Note that \( g \) is strictly convex in \( p \) for each \( x \) . Since \( \left| {H\left( {x, p}\right) }\right| \leq K\parallel p\parallel \) , where \( K \) is a uniform bound for \( F\left( x\right) \), it is easy to see that for all \( x \) in some neighborhood of \( u\left( t\right) \), the function \( p \mapsto g\left( {x, p}\right) \) attains a unique minimum at \( p = p\left( x\right) \), and that we have (for all \( x \) near \( u\left( t\right) \) ) \( \parallel p\left( x\right) \parallel \leq c \) for some constant \( c \) . We claim further that when \( x = u\left( t\right) \), the \( p \) at which said minimum occurs is \( \xi \left( t\right) \) . Since \[ g\left( {u\left( t\right) ,\xi \left( t\right) }\right) = H\left( {u\left( t\right) ,\xi \left( t\right) }\right) = \langle \xi \left( t\right), v\left( t\right) \rangle \] by (3), this claim follows from the following calculation: \[ \mathop{\min }\limits_{p}\{ g\left( {u\left( t\right), p}\right) - \langle \xi \left( t\right), v\left( t\right) \rangle \} \] \[ = \mathop{\min }\limits_{p}\left\{ {H\left( {u\left( t\right), p}\right) -\langle p, v\left( t\right) \rangle + \frac{\parallel \xi \left( t\right) - p{\parallel }^{2}}{4\sigma }}\right\} \] \[ = \mathop{\min }\limits_{p}\mathop{\max }\limits_{{{v}^{\prime } \in F\left( {u\left( t\right) }\right) }}\left\{ {\left\langle {p,{v}^{\prime } - v\left( t\right) }\right\rangle + \frac{\parallel \xi \left( t\right) - p{\parallel }^{2}}{4\sigma }}\right\} \] \[ = \mathop{\max }\limits_{{{v}^{\prime } \in F\left( {u\left( t\right) }\right) }}\mathop{\min }\limits_{p}\left\{ {\left\langle {p,{v}^{\prime } - v\left( t\right) }\right\rangle + \frac{\parallel \xi \left( t\right) - p{\parallel }^{2}}{4\sigma }}\right\} \] (by the Minimax Theorem) \[ = \mathop{\max }\limits_{{{v}^{\prime } \in F\left( {u\left( t\right) }\right) }}\left\{ {\left\langle {\xi \left( t\right) ,{v}^{\prime } - v\left( t\right) }\right\rangle - \sigma {\begin{Vmatrix}{v}^{\prime } - v\left( t\right) \end{Vmatrix}}^{2}}\right\} = 0\;\left( {\text{ by }\left( 2\right) }\right) . \] The preceding facts concerning \( g \) will allow us to complete the proof of Lemma 9.4; we record here in generic terms the max-min principle involved for its independent interest. 9.5. Lemma. Let \( g\left( {x, p}\right) \) be a locally Lipschitz function such that for each \( x \) in a neighborhood of \( \bar{x} \), the function \( p \mapsto g\left( {x, p}\right) \) has a unique minimum at \( p = p\left( x\right) \), where for some \( c \) we have \( \parallel p\left( x\right) \parallel \leq c \) for all \( x \) near \( \bar{x} \) . Let \( p\left( \bar{x}\right) = \bar{p} \), and suppose that the function \( x \mapsto \mathop{\min }\limits_{p}g\left( {x, p}\right) \) has a local maximum at \( x = \bar{x} \) . Then \( \left( {0,0}\right) \in {\partial }_{C}g\left( {\bar{x},\bar{p}}\right) \) . Proof. Take any \( \left( {y, q}\right) \), together with a positive sequence \( {\lambda }_{i} \) decreasing to 0 . Let \( {p}_{i} = p\left( {\bar{x} - {\lambda }_{i}y}\right) \), the minimizer of \( g\left( {\bar{x} - {\lambda }_{i}y, \cdot }\right) \) . Then \( \begin{Vmatrix}{p}_{i}\end{Vmatrix} \leq c \) for all \( i \) sufficiently large, and by passing to a subsequence we can suppose that \( {p}_{i} \rightarrow {p}_{0} \) for some \( {p}_{0} \) . We have, for any \( p \) , \[ g\left( {\bar{x} - {\lambda }_{i}y, p}\right) \geq g\left( {\bar{x} - {\lambda }_{i}y,{p}_{i}}\right) \] whence, in the limit, \( g\left( {\bar{x}, p}\right) \geq g\left( {\bar{x},{p}_{0}}\right) \) . By uniqueness of the minimizer, we deduce \( {p}_{0} = p\left( \bar{x}\right) = \bar{p} \) . Now consider \[ g\left( {\bar{x} - {\lambda }_{i}y,{p}_{i}}\right) = \mathop{\min }\limits_{p}g\left( {\bar{x} - {\lambda }_{i}y, p}\right) \] \[ \leq \mathop{\min }\limits_{p}g\left( {\bar{x}, \cdot }\right) \;\text{(by the maximizing property of}\bar{x}\text{)} \] \[ \leq g\left( {\bar{x},{p}_{i} + {\lambda }_{i}q}\right) \] Hence \[ g\left( {\bar{x},{p}_{i} + {\lambda }_{i}q}\right) - g\left( {\bar{x} - {\lambda }_{i}y,{p}_{i}}\right) \geq 0. \] Dividing this by \( {\lambda }_{i} \) and taking upper limits gives: \[ {g}^{ \circ }\left( {\bar{x},\bar{p};y, q}\right) \geq \mathop{\limsup }\limits_{{i \rightarrow \infty }}\frac{g\left( {\bar{x},{p}_{i} + {\lambda }_{i}q}\right) - g\left( {\bar{x} - {\lambda }_{i}y,{p}_{i}}\right) }{{\lambda }_{i}} \geq 0. \] Since \( \left( {y, q}\right) \) is arbitrary, we obtain \( \left( {0,0}\right) \in {\partial }_{C}g\left( {\bar{x},\bar{p}}\right) \) as claimed, proving Lemma 9.5. We wish to apply Lemma 9.5 t
1012_(GTM178)Nonsmooth Analysis and Control Theory
84
ar{x}\right) = \bar{p} \) . Now consider \[ g\left( {\bar{x} - {\lambda }_{i}y,{p}_{i}}\right) = \mathop{\min }\limits_{p}g\left( {\bar{x} - {\lambda }_{i}y, p}\right) \] \[ \leq \mathop{\min }\limits_{p}g\left( {\bar{x}, \cdot }\right) \;\text{(by the maximizing property of}\bar{x}\text{)} \] \[ \leq g\left( {\bar{x},{p}_{i} + {\lambda }_{i}q}\right) \] Hence \[ g\left( {\bar{x},{p}_{i} + {\lambda }_{i}q}\right) - g\left( {\bar{x} - {\lambda }_{i}y,{p}_{i}}\right) \geq 0. \] Dividing this by \( {\lambda }_{i} \) and taking upper limits gives: \[ {g}^{ \circ }\left( {\bar{x},\bar{p};y, q}\right) \geq \mathop{\limsup }\limits_{{i \rightarrow \infty }}\frac{g\left( {\bar{x},{p}_{i} + {\lambda }_{i}q}\right) - g\left( {\bar{x} - {\lambda }_{i}y,{p}_{i}}\right) }{{\lambda }_{i}} \geq 0. \] Since \( \left( {y, q}\right) \) is arbitrary, we obtain \( \left( {0,0}\right) \in {\partial }_{C}g\left( {\bar{x},\bar{p}}\right) \) as claimed, proving Lemma 9.5. We wish to apply Lemma 9.5 to our particular function \( g \), at the point \( \bar{x} = u\left( t\right) ,\bar{p} = \xi \left( t\right) \) . To do so, it suffices to check that the function \( x \mapsto \) \( \mathop{\min }\limits_{p}g\left( {x, p}\right) \) has a maximum at \( x = u\left( t\right) \) . We calculate \[ \mathop{\min }\limits_{p}g\left( {x, p}\right) \leq \langle \zeta \left( t\right), x - u\left( t\right) \rangle - \sigma \parallel x - u\left( t\right) {\parallel }^{2} + H\left( {x,\xi \left( t\right) }\right) \] \[ = \langle \zeta \left( t\right), x - u\left( t\right) \rangle - \sigma \parallel x - u\left( t\right) {\parallel }^{2} + \left\langle {\xi \left( t\right) ,{v}^{\prime }}\right\rangle \] (for some \( {v}^{\prime } \in F\left( x\right) \) ) \[ \leq \langle \xi \left( t\right), v\left( t\right) \rangle = g\left( {u\left( t\right) ,\xi \left( t\right) }\right) = \mathop{\min }\limits_{p}g\left( {u\left( t\right), p}\right) , \] which gives the required conclusion. The application of Lemma 9.5 gives \( \left( {0,0}\right) \in {\partial }_{C}g\left( {u\left( t\right) ,\xi \left( t\right) }\right) \), which reduces precisely to the conclusion of Lemma 9.4, which is therefore proven. We return now to characterizing \( {N}_{S}^{P} \) . 9.6. Lemma. Let \( \zeta \in {N}_{S}^{P}\left( {v}_{0}\right) \) . Then there is an arc \( q \) on \( \left\lbrack {0,1}\right\rbrack \) with \( q\left( 1\right) = \) 0 such that \[ \left( {-\dot{q},{v}_{0}}\right) \in {\partial }_{C}H\left( {{u}_{0}, q + \zeta }\right) \text{a.e.,} \] where \( {u}_{0}\left( t\right) \mathrel{\text{:=}} {\int }_{0}^{t}{v}_{0}\left( s\right) {ds} \) . Proof. Observe first that for some \( {\sigma }_{0} > 0 \), the following function \( \varphi \) of \( v \) is minimized over \( S \) at \( v = {v}_{0} \) : \[ \varphi \left( v\right) \mathrel{\text{:=}} \langle - \zeta, v\rangle + {\sigma }_{0}{\begin{Vmatrix}v - {v}_{0}\end{Vmatrix}}^{2}. \] We can assume (by increasing \( {\sigma }_{0} \) if necessary) that \( {v}_{0} \) is the unique minimizer of \( \varphi \) on \( S \) . We introduce the value function \( V : X \rightarrow \left( {-\infty ,\infty }\right) \) of a certain minimization problem on \( X \times X \) : \[ V\left( \alpha \right) \mathrel{\text{:=}} \inf \left\{ {\varphi \left( v\right) : \left( {u, v}\right) \in C, u\left( t\right) - {\int }_{0}^{t}v\left( s\right) {ds} = \alpha \left( t\right) \text{ a.e. }}\right\} . \] 9.7. Exercise. (a) \( V\left( 0\right) = \varphi \left( {v}_{0}\right) \) . (b) The infimum defining \( V\left( \alpha \right) \) is attained. (c) If \( {\alpha }_{i} \rightarrow 0 \) in \( X \) and \( V\left( {\alpha }_{i}\right) \rightarrow V\left( 0\right) \), and if \( \left( {{u}_{i},{v}_{i}}\right) \) is a solution of the problem defining \( V\left( {\alpha }_{i}\right) \), then there is a subsequence of \( \left\{ \left( {{u}_{i},{v}_{i}}\right) \right\} \) converging in \( X \times X \) to \( \left( {{u}_{0},{v}_{0}}\right) \), where \[ {u}_{0}\left( t\right) \mathrel{\text{:=}} {\int }_{0}^{t}{v}_{0}\left( s\right) {ds} \] We now take a subsequence \( {\alpha }_{i} \rightarrow 0 \) for which \( \left( {{u}_{i},{v}_{i}}\right) \rightarrow \left( {{u}_{0},{v}_{0}}\right) \) as in the exercise above, and for which \( {\partial }_{P}V\left( {\alpha }_{i}\right) \) contains an element \( {\zeta }_{i} \) (we have invoked the Proximal Density Theorem). As in \( §{3.1} \), the proximal subgradient inequality for \( V \) at \( {\alpha }_{i} \) translates immediately to the following assertion: the function \( {f}_{i}\left( {u, v}\right) \) has a local minimum over \( C \) at \( \left( {u, v}\right) = \left( {{u}_{i},{v}_{i}}\right) \), where \[ {f}_{i}\left( {u, v}\right) \mathrel{\text{:=}} \langle - \zeta, v\rangle + {\sigma }_{0}{\begin{Vmatrix}v - {v}_{0}\end{Vmatrix}}^{2} - \left\langle {{\zeta }_{i}, u - {\int }_{0}^{t}v}\right\rangle + {\sigma }_{i}{\begin{Vmatrix}u - {\int }_{0}^{t}v - {u}_{i} + {\int }_{0}^{t}{v}_{i}\end{Vmatrix}}^{2}. \] 9.8. Exercise. \( {f}_{i} \) is \( {C}^{2} \), with \[ {f}_{i}^{\prime }\left( {{u}_{i},{v}_{i}}\right) = \left( {-{\zeta }_{i}, - \zeta + 2{\sigma }_{0}\left( {{v}_{i} - {v}_{0}}\right) + {\int }_{t}^{1}{\zeta }_{i}}\right) . \] In light of this exercise, the necessary condition \[ - {f}_{i}^{\prime }\left( {{u}_{i},{v}_{i}}\right) \in {N}_{C}^{P}\left( {{u}_{i},{v}_{i}}\right) \] becomes \[ \left( {{\zeta }_{i},\zeta - 2{\sigma }_{0}\left( {{v}_{i} - {v}_{0}}\right) - {\int }_{t}^{1}{\zeta }_{i}}\right) \in {N}_{C}^{P}\left( {{u}_{i},{v}_{i}}\right) . \] Calling upon Lemma 9.4, this gives \[ \left( {-{\zeta }_{i},{v}_{i}}\right) \in {\partial }_{C}H\left( {{u}_{i},\zeta - 2{\sigma }_{0}\left( {{v}_{i} - {v}_{0}}\right) - {\int }_{t}^{1}{\zeta }_{i}}\right) \text{ a.e. } \] Let us relabel as follows: \[ {q}_{i}\left( t\right) = - {\int }_{t}^{1}{\zeta }_{i} \] Then \[ \left( {-{\dot{q}}_{i},{v}_{i}}\right) \in {\partial }_{C}H\left( {{u}_{i},\zeta + {q}_{i} - 2{\sigma }_{0}\left( {{v}_{i} - {v}_{0}}\right) }\right) \text{ a.e. } \] (4) Since \( F \) is globally Lipschitz of rank \( K \), this implies \[ \begin{Vmatrix}{{\dot{q}}_{i}\left( t\right) }\end{Vmatrix} \leq K\begin{Vmatrix}{\zeta + {q}_{i}\left( t\right) - 2{\sigma }_{0}\left( {{v}_{i}\left( t\right) - {v}_{0}\left( t\right) }\right) }\end{Vmatrix}\text{ a.e.,} \] and Gronwall’s Lemma leads to a uniform bound on \( {\begin{Vmatrix}{q}_{i}\end{Vmatrix}}_{\infty } \), and so to a uniform bound on \( {\begin{Vmatrix}{\dot{q}}_{i}\end{Vmatrix}}_{2} \) . Taking a subsequence to arrange the weak convergence of \( {\dot{q}}_{i} \), we are then able to pass to the limit in (4) (with the help of Theorem 3.5.24) to obtain the conclusion of Lemma 9.6. We are ready now to return to the proof of Theorem 9.1, picking up the thread of the argument at the point (1). We recall from Exercise 3.5.20 that an element \( \zeta \) of \( {\partial }_{P}\widetilde{\ell }\left( {v}_{1}\right) \) is of the form \[ \zeta \left( t\right) = {\zeta }_{0} \in {\partial }_{L}\ell \left( {{\int }_{0}^{1}{v}_{1}\left( s\right) {ds}}\right) . \] Combining this with Lemma 9.6, we deduce from (1) that for some such \( {\zeta }_{0} \), for some element \( w \) of \( X \) having \( \parallel w\parallel \leq \varepsilon \), and for some arc \( q \) having \( q\left( 1\right) = 0 \), we have \[ \left( {-\dot{q},{v}_{2}}\right) \in {\partial }_{C}H\left( {{u}_{2}, - {\zeta }_{0} + q + w}\right) \text{ a.e.,} \] where \( {u}_{2}\left( t\right) \mathrel{\text{:=}} {\int }_{0}^{t}{v}_{2}\left( s\right) {ds} \) . Let us now observe this conclusion for a sequence \( {\varepsilon }_{i} \downarrow 0 \), and label the corresponding \( \left( {{u}_{2},{v}_{2}, q,{\zeta }_{0}, w}\right) \) in the form \( \left( {{u}_{i},{v}_{i},{q}_{i},{\zeta }_{i},{w}_{i}}\right) \) . We have \[ \left( {-{\dot{q}}_{i},{v}_{i}}\right) \in {\partial }_{C}H\left( {{u}_{i}, - {\zeta }_{i} + {q}_{i} + {w}_{i}}\right) \text{ a.e.,} \] where \( {w}_{i} \rightarrow 0,{v}_{i} \rightarrow \bar{v},{u}_{i} \rightarrow \bar{x} \), and where \[ {\zeta }_{i} \in {\partial }_{L}\ell \left( {{\int }_{0}^{1}{v}_{i}\left( s\right) {ds}}\right) \] Let us set \( {p}_{i}\left( t\right) \mathrel{\text{:=}} {q}_{i}\left( t\right) - {\zeta }_{i} \) . Then another straightforward use of Gronwall’s Lemma and the Sequential Compactness Theorem 3.5.24 shows that a subsequence of \( \left\{ {p}_{i}\right\} \) converges uniformly to an arc \( p \) satisfying the conclusions of Theorem 9.1, whose proof is now complete. Remark. The Hamiltonian inclusion incorporates the equality, or maximum principle \[ H\left( {x\left( t\right), p\left( t\right) }\right) = \langle p\left( t\right) ,\dot{x}\left( t\right) \rangle \text{ a.e.,} \] (5) as a consequence of Exercise 9.2. 9.9. Exercise. Let \( F\left( x\right) = \{ {Ax} + {Bu} : u \in C\} \), where \( C \) is a compact convex subset of \( {\mathbb{R}}^{m} \), and where \( A, B \) are \( n \times n \) and \( n \times m \) matrices, respectively. (a) Prove that \( F \) satisfies the Standing Hypotheses and is globally Lipschitz. (b) Let \( \bar{x} \) be a trajectory of \( F \) on \( \left\lbrack {0,1}\right\rbrack \), and let \( \bar{u} \) be a control function realizing \( \bar{x} \) ; i.e., such that \( \bar{u}\left( t\right) \in C \) and \( \dot{\bar{x}}\left( t\right) = A\bar{x}\left( t\right) + \) \( B\bar{u}\left( t\right) \) a.e. Show that the Hamiltonian inclusion of Theorem 9.1 is equivalent to \[ - \dot{p}\left( t\right) = {A}^{ * }p\left( t\right) ,\;\mathop{\max }\limits_{{u \in C}}\langle p\left( t\right) ,{Bu}\rangle = \langle p\left( t\right), B\bar{u}\left( t\right) \rangle \text{ a.e. } \] (This is the linear case of Pontryagin's Maximum Principle.) (c) Observe that \( H \) is convex, and so admits one-sided directional derivatives. (d) Prove that any pair of arcs \( \left( {x, p}\right) \) satisfying the Hamiltonia
1012_(GTM178)Nonsmooth Analysis and Control Theory
85
\), and where \( A, B \) are \( n \times n \) and \( n \times m \) matrices, respectively. (a) Prove that \( F \) satisfies the Standing Hypotheses and is globally Lipschitz. (b) Let \( \bar{x} \) be a trajectory of \( F \) on \( \left\lbrack {0,1}\right\rbrack \), and let \( \bar{u} \) be a control function realizing \( \bar{x} \) ; i.e., such that \( \bar{u}\left( t\right) \in C \) and \( \dot{\bar{x}}\left( t\right) = A\bar{x}\left( t\right) + \) \( B\bar{u}\left( t\right) \) a.e. Show that the Hamiltonian inclusion of Theorem 9.1 is equivalent to \[ - \dot{p}\left( t\right) = {A}^{ * }p\left( t\right) ,\;\mathop{\max }\limits_{{u \in C}}\langle p\left( t\right) ,{Bu}\rangle = \langle p\left( t\right), B\bar{u}\left( t\right) \rangle \text{ a.e. } \] (This is the linear case of Pontryagin's Maximum Principle.) (c) Observe that \( H \) is convex, and so admits one-sided directional derivatives. (d) Prove that any pair of arcs \( \left( {x, p}\right) \) satisfying the Hamiltonian inclusion is such that \( t \mapsto H\left( {x\left( t\right), p\left( t\right) }\right) \) is constant on the underlying interval. The following technical extension of the theorem will be of use to us later. 9.10. Exercise. Consider again the context of Theorem 9.1, but extended as follows: for some \( r \geq 0 \), for some \( z\left( \cdot \right) \in {L}_{n}^{2}\left\lbrack {a, b}\right\rbrack \), the trajectory \( \bar{x} \) solves locally the problem of minimizing \[ \ell \left( {x\left( b\right) }\right) + r{\int }_{a}^{b}\parallel x\left( t\right) - z\left( t\right) {\parallel }^{2}{dt} \] relative to the trajectories of \( F \) originating at \( {x}_{0} \) . Adapt the proof of Theorem 9.1 to deduce the same conclusions with one change: the Hamiltonian inclusion becomes \[ \left( {-\dot{p}\left( t\right) ,\dot{\bar{x}}\left( t\right) }\right) \in {\partial }_{C}H\left( {\bar{x}\left( t\right), p\left( t\right) }\right) - {2r}\left( {\bar{x}\left( t\right) - z\left( t\right) ,0}\right) \text{ a.e. } \] (Hint. The derivative of the functional \[ x \mapsto {\int }_{a}^{b}\parallel x\left( t\right) - z\left( t\right) {\parallel }^{2}{dt} = \parallel x - z{\parallel }_{2}^{2} \] is apparent.) ## The Case of Terminal Constraints We consider now the optimal control problem in the presence of explicit constraints on \( x\left( b\right) \) . Specifically, we examine the problem of minimizing \( \ell \left( {x\left( b\right) }\right) \) over the trajectories \( x \) of \( F \) on \( \left\lbrack {a, b}\right\rbrack \) which satisfy \[ x\left( a\right) = {x}_{0},\;x\left( b\right) \in S. \] Here, \( S \) is a given closed subset of \( {\mathbb{R}}^{n} \), and the hypotheses on the rest of the data are unchanged. The proof of the following result uses value function analysis to derive the necessary conditions for the terminally constrained problem from those previously derived for the free endpoint case. 9.11. Theorem. Let \( x \) solve locally the optimal control problem described above. Then there exists an arc \( p \) on \( \left\lbrack {a, b}\right\rbrack \) and a scalar \( {\lambda }_{0} = 0 \) or 1 such that \( {\lambda }_{0} + \parallel p\left( t\right) \parallel \neq 0\forall t \in \left\lbrack {a, b}\right\rbrack \), and such that \[ \left( {-\dot{p}\left( t\right) ,\dot{x}\left( t\right) }\right) \in {\partial }_{C}H\left( {x\left( t\right), p\left( t\right) }\right) \text{ a.e.,}\;a \leq t \leq b, \] \[ - p\left( b\right) \in {\lambda }_{0}{\partial }_{L}\ell \left( {x\left( b\right) }\right) + {N}_{S}^{L}\left( {x\left( b\right) }\right) . \] Proof. We take \( \left\lbrack {a, b}\right\rbrack = \left\lbrack {0,1}\right\rbrack ,{x}_{0} = 0 \), and denote the solution by \( \bar{x} \) . For \( \alpha \in {\mathbb{R}}^{n} \), consider the problem \( P\left( \alpha \right) \) of minimizing \[ \ell \left( {x\left( 1\right) }\right) + {\int }_{0}^{1}\parallel x\left( t\right) - \bar{x}\left( t\right) {\parallel }^{2}{dt} \] over the trajectories of \( F \) satisfying \[ x\left( 0\right) = 0,\;x\left( 1\right) \in S + \alpha ,\;\parallel x - \bar{x}{\parallel }_{\infty } \leq {\varepsilon }_{0}. \] Evidently, \( \bar{x} \) is the unique solution of \( P\left( 0\right) \) . We designate by \( V\left( \alpha \right) \in \) \( ( - \infty ,\infty \rbrack \) the value of the problem \( P\left( \alpha \right) \) . Sequential compactness of trajectories provides a few salient properties of \( V \) : ## 9.12. Exercise. (a) Whenever \( V\left( \alpha \right) < \infty \), the problem \( P\left( \alpha \right) \) admits a solution. (b) If \( {x}_{i} \) solves \( P\left( {\alpha }_{i}\right) \), where \( {\alpha }_{i} \rightarrow 0 \) and \( {x}_{i} \) converges uniformly to an arc \( x \), then \( x = \bar{x} \) . (c) \( V \) is lower semicontinuous. Suppose that \( V \) admits a proximal subgradient \( {\zeta }_{\alpha } \) at \( \alpha \in \operatorname{dom}V \) ; then, for all \( {\alpha }^{\prime } \) sufficiently near \( \alpha \), we have \[ V\left( {\alpha }^{\prime }\right) - V\left( \alpha \right) + \sigma {\begin{Vmatrix}{\alpha }^{\prime } - \alpha \end{Vmatrix}}^{2} \geq \left\langle {{\zeta }_{\alpha },{\alpha }^{\prime } - \alpha }\right\rangle . \] Let \( {x}_{\alpha } \) solve \( P\left( \alpha \right) \), and let \( {c}_{\alpha } \in S \) satisfy \[ {x}_{\alpha }\left( 1\right) = {c}_{\alpha } + \alpha \] In light of Exercise 9.12, we can suppose that \( {\begin{Vmatrix}{x}_{\alpha } - \bar{x}\end{Vmatrix}}_{\infty } < {\varepsilon }_{0} \) when \( \alpha \) is small. For any trajectory \( x \) originating at 0 and for any \( c \in S \), we have \( x\left( 1\right) \in \) \( S + \left( {x\left( 1\right) - c}\right) \), whence, if \( \parallel x - \bar{x}{\parallel }_{\infty } < {\varepsilon }_{0} \) , \[ \ell \left( {x\left( 1\right) }\right) + {\int }_{0}^{1}\parallel x\left( t\right) - \bar{x}\left( t\right) {\parallel }^{2}{dt} \geq V\left( {x\left( 1\right) - c}\right) . \] If \( {\begin{Vmatrix}x - {x}_{\alpha }\end{Vmatrix}}_{\infty } \) and \( \begin{Vmatrix}{c - {c}_{\alpha }}\end{Vmatrix} \) are small enough, this allows us to set \( {\alpha }^{\prime } = x\left( 1\right) - c \) in the preceding inequality and combine it with the last one to deduce that the functional \[ \ell \left( {x\left( 1\right) }\right) + {\int }_{0}^{1}\parallel x\left( t\right) - \bar{x}\left( t\right) {\parallel }^{2}{dt} - \left\langle {{\zeta }_{\alpha }, x\left( 1\right) - c}\right\rangle + \sigma {\begin{Vmatrix}x\left( 1\right) - {x}_{\alpha }\left( 1\right) - c + {c}_{\alpha }\end{Vmatrix}}^{2} \] is minimized locally by \( x = {x}_{\alpha }, c = {c}_{\alpha } \) (relative to trajectories \( x \) of \( F \) originating at 0 and points \( c \in S \) ). Setting \( c = {c}_{\alpha } \), we observe that \( {x}_{\alpha } \) solves locally the problem of minimizing over trajectories \( x \) originating at 0 the cost functional \[ \ell \left( {x\left( 1\right) }\right) - \left\langle {{\zeta }_{\alpha }, x\left( 1\right) }\right\rangle + \sigma {\begin{Vmatrix}x\left( 1\right) - {x}_{\alpha }\left( 1\right) \end{Vmatrix}}^{2} + {\int }_{0}^{1}\parallel x\left( t\right) - \bar{x}\left( t\right) {\parallel }^{2}{dt} \] subject to no terminal constraint. Setting \( x = {x}_{\alpha } \) leads to the conclusion that for all \( c \in S \) sufficiently near \( {c}_{\alpha } \) we have \[ \left\langle {{\zeta }_{\alpha }, c}\right\rangle + \sigma {\begin{Vmatrix}c - {c}_{\alpha }\end{Vmatrix}}^{2} \geq \left\langle {{\zeta }_{\alpha },{c}_{\alpha }}\right\rangle \] this implies \( - {\zeta }_{\alpha } \in {N}_{S}^{P}\left( {c}_{\alpha }\right) \) . Applying Exercise 9.10 to the free endpoint problem solved by \( {x}_{\alpha } \), we deduce the existence of an arc \( {p}_{\alpha } \) such that \[ \left( {-{\dot{p}}_{\alpha },{\dot{x}}_{\alpha }}\right) \in {\partial }_{C}H\left( {{x}_{\alpha },{p}_{\alpha }}\right) - 2\left( {{x}_{\alpha }\left( t\right) - \bar{x}\left( t\right) ,0}\right) \text{ a.e. }, \] (6) \[ - {p}_{\alpha }\left( 1\right) + {\zeta }_{\alpha } \in {\partial }_{L}\ell \left( {{x}_{\alpha }\left( 1\right) }\right) \] (7) \[ - {\zeta }_{\alpha } \in {N}_{S}^{P}\left( {c}_{\alpha }\right) \] (8) We now consider a sequence \( {\alpha }_{i} \rightarrow 0 \) admitting \( {\zeta }_{i} \in {\partial }_{P}V\left( {\alpha }_{i}\right) \) ; this is possible by the Proximal Density Theorem 1.3.1. Denote the \( {x}_{{\alpha }_{i}},{p}_{{\alpha }_{i}},{c}_{{\alpha }_{i}} \) above by the simpler notation \( {x}_{i},{p}_{i},{c}_{i} \) . We may suppose in light of Exercise 9.12 that the corresponding solutions \( {x}_{i} \) converge uniformly to \( \bar{x} \), by passing to a subsequence. Note that \( {c}_{i} = {x}_{i}\left( 1\right) - {\alpha }_{i} \) converges to \( \bar{x}\left( 1\right) \) . The end of the proof consists of passing to the limit in (6), (7), and (8), and there are two cases depending on whether or not the sequence \( {\zeta }_{i} \) is bounded. Consider first the case in which \( {\zeta }_{i} \) is bounded. Then we may suppose that \( {\zeta }_{i} \rightarrow \zeta \in {N}_{S}^{L}\left( {\bar{x}\left( 1\right) }\right) \), and an application of the Sequential Compactness Theorem 3.5.24 leads to an arc \( p \) satisfying \[ \left( {-\dot{p},\dot{\bar{x}}}\right) \in {\partial }_{C}H\left( {\bar{x}, p}\right) ,\; - p\left( 1\right) \in {\partial }_{L}\ell \left( {\bar{x}\left( 1\right) }\right) + {N}_{S}^{L}\left( {\bar{x}\left( 1\right) }\right) , \] that is, the conclusions of the theorem with \( {\lambda }_{0} = 1 \) . Now suppose that \( {\zeta }_{i} \) is unbounded, and pass to a subsequence to arrange \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \rightarrow \infty \) . Dividing by \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \) in (6) and (7) and setting \( {q}_{i} \mathrel{\text{:=}} {p}_{i}/\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \), we obtain (Exercise 9.2): \[
1012_(GTM178)Nonsmooth Analysis and Control Theory
86
her or not the sequence \( {\zeta }_{i} \) is bounded. Consider first the case in which \( {\zeta }_{i} \) is bounded. Then we may suppose that \( {\zeta }_{i} \rightarrow \zeta \in {N}_{S}^{L}\left( {\bar{x}\left( 1\right) }\right) \), and an application of the Sequential Compactness Theorem 3.5.24 leads to an arc \( p \) satisfying \[ \left( {-\dot{p},\dot{\bar{x}}}\right) \in {\partial }_{C}H\left( {\bar{x}, p}\right) ,\; - p\left( 1\right) \in {\partial }_{L}\ell \left( {\bar{x}\left( 1\right) }\right) + {N}_{S}^{L}\left( {\bar{x}\left( 1\right) }\right) , \] that is, the conclusions of the theorem with \( {\lambda }_{0} = 1 \) . Now suppose that \( {\zeta }_{i} \) is unbounded, and pass to a subsequence to arrange \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \rightarrow \infty \) . Dividing by \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \) in (6) and (7) and setting \( {q}_{i} \mathrel{\text{:=}} {p}_{i}/\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \), we obtain (Exercise 9.2): \[ \left( {-{\dot{q}}_{i},{\dot{x}}_{i}}\right) \in {\partial }_{C}H\left( {{x}_{i},{q}_{i}}\right) - 2{\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix}}^{-1}\left( {{x}_{\alpha }\left( t\right) - \bar{x}\left( t\right) ,0}\right) , \] \[ - {q}_{i}\left( 1\right) \in {\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix}}^{-1}{\partial }_{L}\ell \left( {{x}_{i}\left( 1\right) }\right) - {\zeta }_{i}/\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \] where \( - {\zeta }_{i}/\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \in {N}_{S}^{P}\left( {c}_{i}\right) \) . We use Gronwall’s Lemma to deduce that \( {\begin{Vmatrix}{q}_{i}\end{Vmatrix}}_{\infty } \) is bounded, following which Theorem 3.5.24 once more applies to give an arc \( p = \lim {q}_{i} \) such that \[ \left( {-\dot{p},\dot{\bar{x}}}\right) \in {\partial }_{C}H\left( {\bar{x}, p}\right) ,\;0 \neq - p\left( 1\right) \in {N}_{S}^{L}\left( {\bar{x}\left( 1\right) }\right) . \] These are the required conclusions with \( {\lambda }_{0} = 0 \), there remaining only to check that \( p\left( t\right) \neq 0\forall t \in \left\lbrack {0,1}\right\rbrack \) . But we have \( \parallel \dot{p}\left( t\right) \parallel \leq K\parallel p\left( t\right) \parallel \), by Exercise 9.2, so that if \( p\left( t\right) = 0 \) for some \( t \), then \( p\left( t\right) = 0 \) everywhere (by Gronwall’s Lemma). Since \( p\left( 1\right) \neq 0 \), this does not occur. Remark. If the solution \( x \) of Theorem 9.11 is such that \( x\left( b\right) \in \operatorname{int}S \) , then \( {\lambda }_{0} = 1 \) necessarily and the necessary conditions become those of Theorem 9.1 for the free endpoint problem. The condition that \( {\lambda }_{0} + \parallel p\left( t\right) \parallel \) be different from 0 is required in Theorem 9.11 to avoid triviality of the necessary conditions, for they always hold with \( {\lambda }_{0} = 0, p \equiv 0 \), when \( x \) is a trajectory. The case \( {\lambda }_{0} = 0 \) is termed abnormal, and can be shown to arise when the constraints are so tight as to make the cost functional irrelevant. ## 9.13. Exercise. (a) Show that \( {\lambda }_{0} = 1 \) in Theorem 9.11 if \( x\left( b\right) \in \operatorname{int}S \) . (b) Show that \( {\lambda }_{0} = 0, p \equiv 0 \) satisfies the Hamiltonian inclusion when \( x \) is any trajectory. (c) Let \( n = 2,\left\lbrack {a, b}\right\rbrack = \left\lbrack {0,1}\right\rbrack ,{x}_{0} = \left( {0,0}\right), F\left( {x, y}\right) \equiv \bar{B}, S = \{ 1\} \times \mathbb{R} \) . Let \( \ell \left( {x, y}\right) = \ell \left( y\right) \) be any function such that \( 0 \notin {\partial }_{L}\ell \left( 0\right) \) . What is the unique trajectory solving the problem of Theorem 9.11 in this case? Show that \( {\lambda }_{0} = 0 \) necessarily. (d) Prove that in the context of Theorem 9.11, when \( {\lambda }_{0} = 0 \), then \( p\left( t\right) = 0 \) for some \( t \) iff \( p \) is identically zero. ## Constancy of the Hamiltonian In the general context of Theorem 9.11, in contrast to the setting of Exercise 9.9, it does not follow from the Hamiltonian inclusion that \( H\left( {x\left( t\right), p\left( t\right) }\right) \) is constant. But this condition can be obtained for at least one arc \( p \), as we now show. 9.14. Theorem. In Theorem 9.11, we may add to the conditions satisfied by \( \left( {x, p}\right) \) the constancy of the function \( t \mapsto H\left( {x\left( t\right), p\left( t\right) }\right) \) on \( \left\lbrack {a, b}\right\rbrack \) . Proof. The proof will consist of "bootstrapping" from Theorem 9.11, by means of a device known as the Erdmann transform. Let us consider \( \left( {n + 1}\right) \) - dimensional arcs which are trajectories of the multifunction \( \widetilde{F} \) defined by \[ \widetilde{F}\left( {{x}^{0}, x}\right) \mathrel{\text{:=}} \left\{ {\left( {u, v}\right) \in \mathbb{R} \times {\mathbb{R}}^{n} : \left| u\right| \leq \alpha, v \in \left( {1 + u}\right) F\left( x\right) }\right\} , \] where \( \alpha \) is a fixed parameter in \( \left( {0,1}\right) \) . We will take \( \left\lbrack {a, b}\right\rbrack = \left\lbrack {0,1}\right\rbrack \), and we consider problem (E) of minimizing \( \ell \left( {x\left( 1\right) }\right) \) over the trajectories \( \left( {{x}^{0}, x}\right) \) of \( \widetilde{F} \) on \( \left\lbrack {0,1}\right\rbrack \) satisfying \[ \left( {{x}^{0}, x}\right) \left( 0\right) = \left( {1,{x}_{0}}\right) ,\;{x}^{0}\left( 1\right) = 1,\;x\left( 1\right) \in S. \] We denote by \( \bar{x} \) the solution of the original problem (P). 9.15. Lemma. For \( \alpha \) sufficiently small, the arc \( \left( {1,\bar{x}}\right) \) is an admissible trajectory of \( \widetilde{F} \) which solves locally problem (E). That \( \left( {1,\bar{x}}\right) \) is an admissible trajectory for (E) is easily observed. Suppose that \( \left( {{x}^{0}, x}\right) \) is another such, and that \( \ell \left( {x\left( 1\right) }\right) < \ell \left( {\bar{x}\left( 1\right) }\right) \) ; we will manufacture a contradiction by exhibiting an arc \( y \) feasible for (P) with \( \ell \left( {y\left( 1\right) }\right) < \ell \left( {\bar{x}\left( 1\right) }\right) \) . For each \( t \in \left\lbrack {0,1}\right\rbrack \), a unique \( \tau \left( t\right) \) in \( \left\lbrack {0,1}\right\rbrack \) is defined by the equation \[ \tau + {x}^{0}\left( \tau \right) - 1 = t \] since the function \( \tau \mapsto \tau + {x}^{0}\left( \tau \right) - 1 \) is strictly monotone increasing, with value 0 at 0 and 1 at 1 . The function \( \tau \left( t\right) \) is Lipschitz by the Inverse Function Theorem 3.3.12. Thus the relation \( y\left( t\right) \mathrel{\text{:=}} x\left( {\tau \left( t\right) }\right) \) defines \( y \) as an arc on \( \left\lbrack {0,1}\right\rbrack \) with \( y\left( 0\right) = {x}_{0} \) and \( y\left( 1\right) = x\left( 1\right) \in S \) . We calculate \[ \frac{d}{dt}y\left( t\right) = \frac{\dot{x}\left( \tau \right) }{1 + {\dot{x}}^{0}\left( \tau \right) } \] from which it follows that \( y \) is a trajectory of \( F \) . Thus \( y \) is feasible for (P); yet, \( \ell \left( {y\left( 1\right) }\right) = \ell \left( {x\left( 1\right) }\right) < \ell \left( {\bar{x}\left( 1\right) }\right) \) . This contradiction completes the proof, except that we have ignored the fact that \( \bar{x} \) may be only a local solution of (P) relative to \( \parallel x - \bar{x}{\parallel }_{\infty } \leq {\varepsilon }_{0} \) . To take account of this, it suffices to observe that taking both \( \left( {{x}^{0}, x}\right) \) uniformly close enough to \( \left( {1,\bar{x}}\right) \) and \( \alpha \) sufficiently small guarantees that \( y \) satisfies \( \parallel y - \bar{x}{\parallel }_{\infty } < {\varepsilon }_{0} \) . The next step in the proof is to apply the necessary conditions of Theorem 9.11 to the solution \( \left( {1,\bar{x}}\right) \) of problem (E). This will involve the Hamiltonian \[ \widetilde{H}\left( {{x}^{0}, x,{p}^{0}, p}\right) \mathrel{\text{:=}} \max \left\{ {{p}^{0}u+\langle p, v\rangle : \left| u\right| \leq \alpha, v \in \left( {1 + u}\right) F\left( x\right) }\right\} \] \[ = \max \left\{ {{p}^{0}u + \left( {1 + u}\right) H\left( {x, p}\right) : \left| u\right| \leq \alpha }\right\} \] \[ = H\left( {x, p}\right) + \alpha \left| {{p}^{0} + H\left( {x, p}\right) }\right| . \] Since this is independent of \( {x}^{0} \), it follows from the Hamiltonian inclusion that \( {p}^{0} \) is constant, and that we have \[ \left( {-\dot{p},\dot{\bar{x}}}\right) \in {\partial }_{C}H\left( {\dot{\bar{x}}, p}\right) + \left( {{\alpha K}\parallel p\parallel \bar{B}}\right) \times \left( {{\alpha K}\bar{B}}\right) , \] (9) where (as in the proof of Theorem 9.1) the global bound and Lipschitz constant \( K \) for \( F \) can be assumed to exist. The Hamiltonian inclusion for \( \widetilde{H} \) also yields a.e. (see (5)): \[ \widetilde{H}\left( {1,\bar{x},{p}^{0}, p}\right) = \left\langle {\left( {0,\dot{\bar{x}}}\right) ,\left( {{p}^{0}, p}\right) }\right\rangle = \langle \dot{\bar{x}}, p\rangle \] \[ \leq H\left( {\bar{x}, p}\right) \leq H\left( {\bar{x}, p}\right) + \alpha \left| {{p}^{0} + H\left( {\bar{x}, p}\right) }\right| = \widetilde{H}\left( {1,\bar{x},{p}^{0}, p}\right) . \] It follows from this that \[ H\left( {\bar{x}\left( t\right), p\left( t\right) }\right) \equiv - {p}^{0}\forall t \in \left\lbrack {0,1}\right\rbrack . \] (10) Note also that the transversality conditions applied to (E) are the following: \[ - \left( {{p}^{0}\left( 1\right), p\left( 1\right) }\right) \in {\lambda }_{0}\left( {0,{\partial }_{L}\ell \left( {\bar{x}\left( 1\right) }\right) }\right) + {N}_{\{ 1\} \times S}^{L}\left( {1,\bar{x}\left( 1\right) }\right) \] which implies \[ - p\left( 1\right) \in {\lambda }_{0}{\partial }_{L}\ell \left( {\bar{x}\left( 1\right) }\right) + {N}_{S}^{L}\left( {\bar{x}\left( 1\right) }\right) . \] (11) Finally, we have the nontriviality condition \( {\lambda }_{0} + \b
1012_(GTM178)Nonsmooth Analysis and Control Theory
87
ght) ,\left( {{p}^{0}, p}\right) }\right\rangle = \langle \dot{\bar{x}}, p\rangle \] \[ \leq H\left( {\bar{x}, p}\right) \leq H\left( {\bar{x}, p}\right) + \alpha \left| {{p}^{0} + H\left( {\bar{x}, p}\right) }\right| = \widetilde{H}\left( {1,\bar{x},{p}^{0}, p}\right) . \] It follows from this that \[ H\left( {\bar{x}\left( t\right), p\left( t\right) }\right) \equiv - {p}^{0}\forall t \in \left\lbrack {0,1}\right\rbrack . \] (10) Note also that the transversality conditions applied to (E) are the following: \[ - \left( {{p}^{0}\left( 1\right), p\left( 1\right) }\right) \in {\lambda }_{0}\left( {0,{\partial }_{L}\ell \left( {\bar{x}\left( 1\right) }\right) }\right) + {N}_{\{ 1\} \times S}^{L}\left( {1,\bar{x}\left( 1\right) }\right) \] which implies \[ - p\left( 1\right) \in {\lambda }_{0}{\partial }_{L}\ell \left( {\bar{x}\left( 1\right) }\right) + {N}_{S}^{L}\left( {\bar{x}\left( 1\right) }\right) . \] (11) Finally, we have the nontriviality condition \( {\lambda }_{0} + \begin{Vmatrix}\left( {{p}^{0}, p}\right) \end{Vmatrix} \neq 0 \) . If it were the case that \( {\lambda }_{0} \) and \( \parallel p\left( t\right) \parallel \) were both 0 for some \( t \), then \( {p}^{0} = 0 \) also (in light of (10)), which cannot be. Thus we conclude \[ {\lambda }_{0} + \parallel p\left( t\right) \parallel \neq 0\forall t \in \left\lbrack {0,1}\right\rbrack \] (12) as required. If we examine (9)-(12) we see exactly the set of conditions that we are attempting to obtain, except for the superfluous term involving \( \alpha \) in (9). The next step is evident: we consider having done all the above for a sequence \( {\alpha }_{i} \downarrow 0 \), and pass to the limit. The corresponding arcs \( {p}_{i} \) satisfy \[ \begin{Vmatrix}{\dot{p}}_{i}\end{Vmatrix} \leq K\left( {1 + {\alpha }_{i}}\right) \begin{Vmatrix}{p}_{i}\end{Vmatrix} \] as a consequence of (9), and so \( {\begin{Vmatrix}{p}_{i}\end{Vmatrix}}_{\infty } \) is bounded iff \( {p}_{i}\left( 1\right) \) is bounded (by Gronwall’s Lemma). In the unbounded case, we arrange \( \begin{Vmatrix}{{p}_{i}\left( 1\right) }\end{Vmatrix} \rightarrow \infty \), and then normalize \( {p}_{i} \) in (9)-(11) by passing to \( {p}_{i}/\begin{Vmatrix}{{p}_{i}\left( 1\right) }\end{Vmatrix} \), much as in the last step of the proof of Theorem 9.11. The required conclusions then emerge in the limit via sequential compactness (and with \( {\lambda }_{0} = 0 \) ). The resulting arc \( p \) has \( \parallel p\left( 1\right) \parallel = 1 \), so \( p \) is never zero (Exercise 9.13(d)) and we have nontriviality. When \( {p}_{i}\left( 1\right) \), and hence \( {\begin{Vmatrix}{p}_{i}\end{Vmatrix}}_{\infty } \) is bounded, sequential compactness can be applied directly to (9)-(11) without renormalizing to yield the required Hamiltonian inclusion, constancy, and transversality, but a danger lurks: triviality; perhaps the sequences \( {p}_{i} \) and \( {\lambda }_{{0}_{i}} \) both converge to 0 . We deal with this by considering two subcases. In the first, \( {\lambda }_{{0}_{i}} = 1 \) infinitely often; in that event the danger does not arise. In the limit, we get \( {\lambda }_{0} = 1 \) . In the second, all \( {\lambda }_{{0}_{i}} = 0 \) beyond a certain point. Then we have \( {p}_{i}\left( 1\right) \neq 0 \) necessarily, in view of (12), and we can renormalize (9)-(11) by dividing by \( \begin{Vmatrix}{{p}_{i}\left( 1\right) }\end{Vmatrix} \) . Then sequential compactness applies to give the required (nontrivial) conclusions in the limit, with \( {\lambda }_{0} = 0 \) . ## Free Time Problems The transformation device introduced in the proof of the theorem can be used to treat free time problems, in which the underlying interval \( \left\lbrack {a, b}\right\rbrack \) is itself a variable. We illustrate this now in a simple setting. 9.16. Exercise. Let \( \bar{x} \) and \( \bar{T} > 0 \) solve the following problem: to minimize \( \ell \left( {T, x\left( T\right) }\right) \) over those trajectories \( x \) on \( \left\lbrack {0, T}\right\rbrack \) originating at \( x \) which satisfy \( \left( {t, x\left( t\right) }\right) \in \Omega \forall t \in \left\lbrack {0, T}\right\rbrack \), where \( \Omega \) is a given open set in \( \mathbb{R} \times {\mathbb{R}}^{n} \) . (Thus a form of merely local optimality is involved; we stress that \( T \) is a choice variable in this problem.) The function \( \ell \) is locally Lipschitz. (a) Referring to the proof of Theorem 9.14, show that the arc \( \left( {\bar{T},\bar{x}}\right) \) minimizes locally, over trajectories \( \left( {{x}^{0}, x}\right) \) of \( \widetilde{F} \) on \( \left\lbrack {0,\bar{T}}\right\rbrack \) originating at \( \left( {\bar{T},{x}_{0}}\right) \), the function \( \ell \left( {{x}^{0}\left( \bar{T}\right), x\left( \bar{T}\right) }\right) \), subject now to no explicit constraint on the value of \( \left( {{x}^{0}\left( \bar{T}\right), x\left( \bar{T}\right) }\right) \) . Observe that this is a situation to which Theorem 9.1 applies. Express the resulting necessary conditions. (b) Show that the necessary conditions of (a) translate directly as follows: for some arc \( p \) on \( \left\lbrack {0,\bar{T}}\right\rbrack \) we have \[ \left( {-\dot{p}\left( t\right) ,\dot{\bar{x}}\left( t\right) }\right) \in {\partial }_{C}H\left( {\bar{x}\left( t\right), p\left( t\right) }\right) \text{ a.e.,}\;0 \leq t \leq \bar{T}, \] \[ H\left( {\bar{x}\left( t\right), p\left( t\right) }\right) = h\left( { = \text{ constant }}\right) ,\;0 \leq t \leq \bar{T}, \] \[ \left( {h, - p\left( \bar{T}\right) }\right) \in {\partial }_{L}\ell \left( {\bar{T},\bar{x}\left( \bar{T}\right) }\right) . \] ## 10 Normality and Controllability We now take up the issue of whether it is possible to reach a given equilibrium point \( {x}^{ * } \) in finite time (from a neighborhood of the point, or globally). Without loss of generality, we take \( {x}^{ * } = 0 \), so we are dealing with null controllability, where it is the case that 0 belongs to \( F\left( 0\right) \) . We assume throughout this section that \( F \) is locally Lipschitz. Recall that if \( F\left( x\right) = \{ f\left( x\right) \} \), then no point different from 0 can reach the equilibrium 0 in finite time. The opposite extreme occurs when \( 0 \in \operatorname{int}F\left( 0\right) \) . 10.1. Exercise. Let \( 0 \in \operatorname{int}F\left( 0\right) \) . Prove that every \( {x}_{0} \) sufficiently near 0 can be steered to the origin in finite time; i.e., admits a trajectory \( x \) on some interval \( \left\lbrack {0, T}\right\rbrack \) with \( x\left( 0\right) = {x}_{0}, x\left( T\right) = 0 \) . We are interested in situations intermediate to the extremes above. It turns out that the Hamiltonian inclusion provides a criterion for local null controllability. We will say that the origin is normal provided that for some \( T > 0 \), the only arc \( p \) on \( \left\lbrack {0, T}\right\rbrack \) which satisfies the two conditions \[ \left( {-\dot{p}\left( t\right) ,0}\right) \in {\partial }_{C}H\left( {0, p\left( t\right) }\right) \text{ a.e.,}\;0 \leq t \leq T, \] \[ H\left( {0, p\left( t\right) }\right) = 0\forall t \in \left\lbrack {0, T}\right\rbrack \] is \( p\left( t\right) \equiv 0 \) . (That \( p \equiv 0 \) does satisfy them is apparent from Exercise 9.2.) 10.2. Exercise. (a) If \( 0 \in \operatorname{int}F\left( 0\right) \), then the origin is normal. (b) Let \( n = 2 \) and \( F\left( {x, y}\right) \mathrel{\text{:=}} \{ \left( {y, u}\right) : - 1 \leq u \leq 1\} \) . Then \( 0 \notin \) int \( F\left( 0\right) \) . Show that the origin is normal. (c) Let \( F\left( x\right) = {Ax} + {BC} \) as in Exercise 9.9, where now \( 0 \in \operatorname{int}C \) . Show that the origin is normal iff the following \( n \times {nm} \) matrix is of maximal rank: \( \left\lbrack {{BAB}{A}^{2}B\ldots {A}^{n - 1}B}\right\rbrack \) . 10.3. Theorem. If the origin is normal, then every point \( {x}_{0} \) sufficiently near it can be steered to it in finite time. Proof. Let \( b > 0 \) be such that no nontrivial arc \( p \) on \( \left\lbrack {0, b}\right\rbrack \) satisfies \( \left( {-\dot{p},0}\right) \in \) \( {\partial }_{C}H\left( {0, p}\right), H\left( {0, p\left( t\right) }\right) = 0 \) on \( \left\lbrack {0, b}\right\rbrack \) . For \( \alpha \in {\mathbb{R}}^{n} \), we define \( V\left( \alpha \right) \in \left\lbrack {0,\infty }\right\rbrack \) to be the infimum of \[ {\left( T - b\right) }^{2} + {\int }_{0}^{T}\parallel x\left( t\right) \parallel {dt} \] over all \( T > 0 \) and arcs \( x \) for \( - F \) on \( \left\lbrack {0, T}\right\rbrack \) originating at 0 which satisfy \( x\left( T\right) = \alpha \) . (Note that \( - F \), not \( F \), is used here.) 10.4. Exercise. (a) \( V\left( 0\right) = 0 \), and for \( \alpha = 0 \) the unique solution to the problem defining \( V\left( 0\right) \) is the arc identically 0 on \( \left\lbrack {0, b}\right\rbrack \) . (b) If \( V\left( \alpha \right) < \infty \), then the infimum defining \( V\left( \alpha \right) \) is attained. (Note: Exercise 1.13 helps in dealing with variable intervals.) (c) \( V \) is lower semicontinuous. (d) If \( {\alpha }_{i} \rightarrow 0 \), and if \( \left( {{T}_{i},{x}_{i}}\right) \) is a solution corresponding to \( {\alpha }_{i} \), then \( {T}_{i} \rightarrow b \) and \( \max \left\{ {\begin{Vmatrix}{{x}_{i}\left( t\right) }\end{Vmatrix} : 0 \leq t \leq {T}_{i}}\right\} \rightarrow 0. \) We will show that \( V \) is Lipschitz (and hence finite) in a neighborhood of 0 . This implies that for every \( {x}_{0} \) near 0, there is a trajectory \( x \) of \( - F \) on an interval \( \left\lbrack {0, T}\right\rbrack \) satisfying \( x\left( 0\right) = 0, x\left( T\right) = {x}_{0} \) . Then, reversing time, we see that \( {x}_{0} \) can be steered to 0 in finite time. To show that \( V \) is Lipschitz near 0, we will prove that its proximal sub-gradients are locally bounded.
1012_(GTM178)Nonsmooth Analysis and Control Theory
88
\( V\left( \alpha \right) < \infty \), then the infimum defining \( V\left( \alpha \right) \) is attained. (Note: Exercise 1.13 helps in dealing with variable intervals.) (c) \( V \) is lower semicontinuous. (d) If \( {\alpha }_{i} \rightarrow 0 \), and if \( \left( {{T}_{i},{x}_{i}}\right) \) is a solution corresponding to \( {\alpha }_{i} \), then \( {T}_{i} \rightarrow b \) and \( \max \left\{ {\begin{Vmatrix}{{x}_{i}\left( t\right) }\end{Vmatrix} : 0 \leq t \leq {T}_{i}}\right\} \rightarrow 0. \) We will show that \( V \) is Lipschitz (and hence finite) in a neighborhood of 0 . This implies that for every \( {x}_{0} \) near 0, there is a trajectory \( x \) of \( - F \) on an interval \( \left\lbrack {0, T}\right\rbrack \) satisfying \( x\left( 0\right) = 0, x\left( T\right) = {x}_{0} \) . Then, reversing time, we see that \( {x}_{0} \) can be steered to 0 in finite time. To show that \( V \) is Lipschitz near 0, we will prove that its proximal sub-gradients are locally bounded. We proceed by supposing the contrary: for some sequence \( {\alpha }_{i} \rightarrow 0 \), and sequence \( {\alpha }_{i} \in {\partial }_{P}V\left( {\alpha }_{i}\right) \), we have \( \begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \rightarrow \infty \) . We will derive a contradiction to the normality hypothesis. Arguing as in the proof of Theorem 9.11, we let \( \left( {{T}_{i},{x}_{i}}\right) \) be a solution corresponding to \( {\alpha }_{i} \), and from \( {\zeta }_{i} \in {\partial }_{P}V\left( {\alpha }_{i}\right) \) we deduce that for some \( {\sigma }_{i} > 0,\left( {{T}_{i},{x}_{i}}\right) \) furnishes a minimum for the problem of minimizing \[ \left\langle {-{\zeta }_{i}, x\left( T\right) }\right\rangle + {\left( T - b\right) }^{2} + {\int }_{0}^{T}\parallel x\left( t\right) \parallel {dt} + {\sigma }_{i}{\begin{Vmatrix}x\left( T\right) - {x}_{i}\left( {T}_{i}\right) \end{Vmatrix}}^{2} \] over trajectories \( x \) of \( - F \) on \( \left\lbrack {0, T}\right\rbrack \) which originate at 0 . (The minimum is local in that \( x\left( T\right) \) must be near \( {x}_{i}\left( {T}_{i}\right) \) .) Because of the integral term in the cost, this is not the type of problem to which our previous results apply. We will fix this by redefining the dynamics so as to "absorb" the integral. We set \[ \widetilde{F}\left( {x, y}\right) \mathrel{\text{:=}} - F\left( x\right) \times \{ \parallel x\parallel \} ,\;{y}_{i}\left( t\right) \mathrel{\text{:=}} {\int }_{0}^{T}\begin{Vmatrix}{{x}_{i}\left( \tau \right) }\end{Vmatrix}{dt}. \] Then \( \left( {{x}_{i},{y}_{i}}\right) \) on \( \left\lbrack {0,{T}_{i}}\right\rbrack \) locally minimizes the free-time cost functional \[ \left\langle {-{\zeta }_{i}, x\left( T\right) }\right\rangle + {\left( T - b\right) }^{2} + y\left( T\right) + {\sigma }_{i}{\begin{Vmatrix}x\left( T\right) - {x}_{i}\left( {T}_{i}\right) \end{Vmatrix}}^{2} \] over the trajectories for \( \widetilde{F} \) originating at \( \left( {0,0}\right) \) . The necessary conditions of Exercise 9.16 are available now. After the dust settles, here is what they give in terms of the original variables: \[ \left( {-{\dot{p}}_{i},{\dot{x}}_{i}}\right) \in {\partial }_{C}\widetilde{h}\left( {{x}_{i},{p}_{i}}\right) + \bar{B} \times \{ 0\} \text{ a.e.,}\;0 \leq t \leq {T}_{i}, \] \[ \widetilde{h}\left( {{x}_{i},{p}_{i}}\right) - \begin{Vmatrix}{{x}_{i}\left( t\right) }\end{Vmatrix} = 2\left( {{T}_{i} - b}\right) \forall t \in \left\lbrack {0,{T}_{i}}\right\rbrack , \] \[ {p}_{i}\left( {T}_{i}\right) = {\zeta }_{i} \] where \( \widetilde{h}\left( {x, p}\right) \mathrel{\text{:=}} H\left( {x, - p}\right) \) appears here because \( - F \) is involved rather than \( F \) . We now replace \( {p}_{i} \) by \( {p}_{i}/\begin{Vmatrix}{\zeta }_{i}\end{Vmatrix} \) in these conditions, and implement the now-familiar passage to the limit. We arrive at a nontrivial arc \( \bar{p} \) satisfying \[ \left( {-\dot{\bar{p}},0}\right) \in {\partial }_{C}\widetilde{h}\left( {0,\bar{p}}\right) \text{ a.e.,}\;0 \leq t \leq b, \] \[ \widetilde{h}\left( {0,\bar{p}\left( t\right) }\right) = 0\forall t \in \left\lbrack {0, b}\right\rbrack \text{.} \] In terms of \( H,\bar{p} \) satisfies \[ \left( {-\dot{\bar{p}},0}\right) \in {\partial }_{C}H\left( {0, - \bar{p}}\right) \text{a.e.,} \] \[ H\left( {0, - \bar{p}\left( t\right) }\right) = 0. \] Now define \( p\left( t\right) \mathrel{\text{:=}} - \bar{p}\left( {b - t}\right) \) ; then on \( \left\lbrack {0, b}\right\rbrack \) we have \[ \left( {-\dot{p},0}\right) \in {\partial }_{C}H\left( {0, p}\right) \text{a.e.,} \] \[ H\left( {0, p\left( t\right) }\right) = 0,\;p ≢ 0. \] This contradicts the normality of the origin. The usual route to a conclusion of global null controllability is to combine global asymptotic controllability with local null controllability, as in the following result that builds upon Theorem 5.5. 10.5. Corollary. Let 0 be a normal equilibrium of \( F \) admitting a Lya-pounov pair \( \left( {Q, W}\right) \) . Then any point in \( \operatorname{dom}Q \) can be steered to 0 in finite time. If \( \operatorname{dom}Q = {\mathbb{R}}^{n} \), then all points can be steered to the origin in finite time. 10.6. Exercise. Prove Corollary 10.5. If a system is globally null controllable, then in fact a Lyapounov pair \( \left( {Q, W}\right) \) does exist. This provides a partial converse to Corollary 10.5: 10.7. Exercise. We define the well-known minimal time function: \( Q\left( \alpha \right) \mathrel{\text{:=}} \inf \{ T \geq 0 \) : some trajectory \( x \) has \( x\left( 0\right) = \alpha, x\left( T\right) = 0\} . \) (As usual, we set \( Q\left( \alpha \right) = \infty \) when no trajectory joins \( \alpha \) to 0 in finite time.) (a) Prove that the infimum is attained if \( Q\left( \alpha \right) < \infty \) . (b) Prove that the system \( \left( {Q\left( x\right) + t,\{ 1\} \times F\left( x\right) }\right) \) is weakly decreasing on the set \( {\mathbb{R}}^{n} \smallsetminus \{ 0\} \) . Call upon Theorem 6.1 to deduce \[ h\left( {x,{\partial }_{P}Q\left( x\right) }\right) \leq - 1\forall x \in {\mathbb{R}}^{n} \smallsetminus \{ 0\} . \] Conclude that setting \( W\left( x\right) = 1 \) for \( x \neq 0 \) and \( W\left( 0\right) = 0 \) leads to a Lyapounov pair \( \left( {Q, W}\right) \) for the origin. Observe that \( \operatorname{dom}Q = {\mathbb{R}}^{n} \) iff the control system is globally null controllable. (c) Prove that in fact \( Q \) satisfies \[ h\left( {x,{\partial }_{P}Q\left( x\right) }\right) = - 1\forall x \in {\mathbb{R}}^{n} \smallsetminus \{ 0\} . \] (A uniqueness theorem for \( Q \) appears among the problems in the next section.) ## 11 Problems on Chapter 4 11.1. Let \( v : \left\lbrack {a, b}\right\rbrack \rightarrow {\mathbb{R}}^{n} \) be a bounded measurable function such that \( v\left( t\right) \in \) \( K \) a.e., where \( K \subseteq {\mathbb{R}}^{n} \) is a given closed convex set. Prove that \[ \frac{1}{b - a}{\int }_{a}^{b}v\left( t\right) {dt} \in K \] 11.2. Consider the ordinary differential system \[ \dot{x} = - x + 2{y}^{3},\;\dot{y} = - x. \] (a) Show that the set \( {E}_{r} \mathrel{\text{:=}} \left\{ {\left( {x, y}\right) : {x}^{2} + {y}^{4} \leq {r}^{2}}\right\} \) is invariant under its flow for all \( r \geq 0 \) . (b) Show that \( r\bar{B} \) is invariant iff \( r = 0 \) or \( 1/\sqrt{2} \) . (c) Show that \( {S}_{r} \mathrel{\text{:=}} \{ \left( {x, y}\right) : \max \left\lbrack {\left| x\right| ,\left| y\right| }\right\rbrack \leq r\} \) is invariant only for \( r = 0 \) . 11.3. We now consider a controlled version of the preceding exercise. Specifically we take \[ \dot{x} = - x + 2{y}^{3} + u,\;\dot{y} = - x + v, \] where \( u \) and \( v \) are measurable control functions satisfying \[ \left| {u\left( t\right) }\right| \leq \delta \text{ and }\left| {v\left( t\right) }\right| \leq \Delta \text{ a.e. } \] Given \( r \geq 0 \), find values of \( \delta \) and \( \Delta \) which ensure that \( {S}_{r} \) is weakly invariant (where \( {S}_{r} \) is as defined in Problem 11.2). 11.4. A multifunction \( F \) on \( {\mathbb{R}}^{n} \) has lower Hamiltonian \[ h\left( {x, p}\right) = \left\{ \begin{array}{ll} - \parallel x\parallel \parallel p\parallel & \text{ if }\parallel x\parallel < 1 \\ - 2\parallel p\parallel & \text{ if }\parallel x\parallel \geq 1 \end{array}\right. \] What is \( F \) ? 11.5. Let \( \left\{ {S}_{\alpha }\right\} \) be a family of subsets of \( {\mathbb{R}}^{n} \), each weakly invariant with respect to \( F \) . Prove that \( \operatorname{cl}\mathop{\bigcup }\limits_{\alpha }{S}_{\alpha } \) is weakly invariant with respect to \( F \) . 11.6. Consider \[ F\left( {x, y}\right) = \left\{ \begin{array}{ll} \left( {1,\frac{1}{\sqrt{2}}}\right) & \text{ if }y > \sqrt{1 + {x}^{2}}, \\ \left( {1,0}\right) & \text{ if }y < \sqrt{1 + {x}^{2}}, \\ \{ 1\} \times \left\lbrack {0,\frac{1}{\sqrt{2}}}\right\rbrack & \text{ if }y = \sqrt{1 + {x}^{2}}. \end{array}\right. \] (a) Show that \( F \) satisfies the Standing Hypotheses. (b) Show that for any selection \( f \) for \( F \), the unique Euler solution on \( \left\lbrack {0,1}\right\rbrack \) to \( \left( {\dot{x},\dot{y}}\right) = f\left( {x, y}\right) \) starting at \( \left( {0,1}\right) \) is one of two linear functions, depending entirely upon \( f\left( {0,1}\right) \) . (c) Show that the curve \( \left( {t,\sqrt{1 + {t}^{2}}}\right) \) does not arise as an Euler solution, but that it is nonetheless a trajectory on \( \left\lbrack {0,1}\right\rbrack \) of the differential inclusion \( \left( {\dot{x},\dot{y}}\right) \in F\left( {x, y}\right) . \) (d) Find all trajectories of the differential inclusion on \( \left\lbrack {0,1}\right\rbrack \) beginning at \( \left( {0,1}\right) \) ; there are infinitely many. 11.7. Consider \[ F\left( {x, y}\right) = \{ \left\lbrack {{xu},{yv}}\right\rbrack : u \geq 0, v \geq 0, u + v = 1\} . \] Confirm tha
1012_(GTM178)Nonsmooth Analysis and Control Theory
89
f }y = \sqrt{1 + {x}^{2}}. \end{array}\right. \] (a) Show that \( F \) satisfies the Standing Hypotheses. (b) Show that for any selection \( f \) for \( F \), the unique Euler solution on \( \left\lbrack {0,1}\right\rbrack \) to \( \left( {\dot{x},\dot{y}}\right) = f\left( {x, y}\right) \) starting at \( \left( {0,1}\right) \) is one of two linear functions, depending entirely upon \( f\left( {0,1}\right) \) . (c) Show that the curve \( \left( {t,\sqrt{1 + {t}^{2}}}\right) \) does not arise as an Euler solution, but that it is nonetheless a trajectory on \( \left\lbrack {0,1}\right\rbrack \) of the differential inclusion \( \left( {\dot{x},\dot{y}}\right) \in F\left( {x, y}\right) . \) (d) Find all trajectories of the differential inclusion on \( \left\lbrack {0,1}\right\rbrack \) beginning at \( \left( {0,1}\right) \) ; there are infinitely many. 11.7. Consider \[ F\left( {x, y}\right) = \{ \left\lbrack {{xu},{yv}}\right\rbrack : u \geq 0, v \geq 0, u + v = 1\} . \] Confirm that the Standing Hypotheses hold, and prove that the attainable set \( \mathcal{A}\left( {\left( {1,1}\right) ;1}\right) \) is not convex. (Hint. Show that while both \( \left( {e,1}\right) \) and \( \left( {1, e}\right) \) are attainable from \( \left( {1,1}\right) \) in time 1, their midpoint is not.) 11.8. Let \( F \) satisfy the Standing Hypotheses, and let \( S \subseteq {\mathbb{R}}^{n} \) be a nonempty, compact, wedged subset of \( {\mathbb{R}}^{n} \) which is weakly avoidable with respect to \( F \) ; that is, \( \operatorname{cl}\left\lbrack {\operatorname{comp}\left( S\right) }\right\rbrack \) is weakly invariant with respect to \( F \) . Assume also that \( \operatorname{cl}\left\lbrack {\operatorname{comp}\left( S\right) }\right\rbrack \) is regular. Then \( S \) contains a zero of \( F \) . 11.9. Let \( g : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{n} \) be a continuous function satisfying the tangency condition \[ \mathop{\lim }\limits_{\substack{{{x}^{\prime } \rightarrow x} \\ {{x}^{\prime } \in S} \\ {\lambda \downarrow 0} }}\frac{{d}_{S}\left\lbrack {{\lambda g}\left( {x}^{\prime }\right) + \left( {1 - \lambda }\right) x}\right\rbrack }{\lambda } = 0 \] where \( S \subseteq {\mathbb{R}}^{n} \) is a compact, wedged, and homeomorphically convex set. (a) Prove that \( g \) has a fixed point in \( S \) . (b) Show that the tangency condition always holds if \( S \) is convex and \( g \) maps \( S \) to itself (the case of Brouwer’s Theorem). 11.10. Prove the following Fixed Point Theorem of Kakutani: let \( C \) be a compact convex subset of \( {\mathbb{R}}^{n} \), and let the multifunction \( G \) satisfy the Standing Hypotheses. Suppose that \[ G\left( x\right) \cap C \neq \varnothing \forall x \in S. \] Then there exists \( \widehat{x} \in C \) such that \( \widehat{x} \in G\left( \widehat{x}\right) \) . 11.11. Let \( F \) be an upper semicontinuous multifunction on \( {\mathbb{R}}^{n} \), with images that are nonempty compact convex subsets of \( {\mathbb{R}}^{n} \) . Prove that for any given \( \varepsilon > 0 \) there exists a locally Lipschitz selection \[ {f}_{\varepsilon }\left( x\right) \in F\left( x\right) + {\varepsilon B}\forall x \in {\mathbb{R}}^{n}. \] 11.12. Consider the linear differential system \[ \dot{x}\left( t\right) = {Ax}\left( t\right) \] where \( A \) is an \( n \times n \) constant real matrix. Prove that the following are equivalent: (a) the nonnegative orthant \( {\mathbb{R}}_{ + }^{n} \) is invariant; (b) \( {e}^{tA}{\mathbb{R}}_{ + }^{n} \subseteq {\mathbb{R}}_{ + }^{n} \) for all \( t \geq 0 \) ; and (c) \( A \) is a Metzler matrix; that is, its off-diagonal entries are nonnegative. 11.13. Let \( S \) be a nonempty compact subset of \( {\mathbb{R}}^{n} \), and let \( f : {\mathbb{R}}^{n} \rightarrow {\mathbb{R}}^{n} \) be continuous. Prove that \( \left( {S,\{ f\} }\right) \) is weakly invariant iff \( f\left( x\right) \in {T}_{S}^{C}\left( x\right) \) \( \forall x \in {\mathbb{R}}^{n} \) . 11.14. (a) Suppose that \( S \) is a compact subset of \( {\mathbb{R}}^{n} \), and that \[ h\left( {x,\zeta }\right) < 0\forall \zeta \in {N}_{S}^{P}\left( x\right) ,\;\zeta \neq 0,\forall x \in S. \] Show by example that \( S \) may nonetheless fail to be locally attainable. (b) Let the compact set \( S \) in \( {\mathbb{R}}^{n} \) be described in the form \[ \left\{ {x \in {\mathbb{R}}^{n} : f\left( x\right) \leq 0}\right\} \] where the locally Lipschitz function \( f : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) satisfies \[ f\left( x\right) = 0 \Rightarrow 0 \notin {\partial }_{L}f\left( x\right) . \] Show that if \[ h\left( {x,\zeta }\right) < 0\forall \zeta \in {\partial }_{L}f\left( x\right) ,\forall x \in S\text{such that}f\left( x\right) = 0\text{,} \] then \( S \) is locally attainable. 11.15. Lyapounov theory can be developed for sets as well as for equilibrium points. Let \( S \) be a nonempty compact subset of \( {\mathbb{R}}^{n} \) . (The case of an equilibrium \( {x}^{ * } \) corresponds to taking \( S = \left\{ {x}^{ * }\right\} \) .) A Lyapounov pair \( \left( {Q, W}\right) \) for \( S \) is defined to be a pair of strictly positive functions \( Q, W \in \mathcal{F}\left( {\operatorname{comp}S}\right) \) such that the sublevel sets \( \{ x : Q\left( x\right) \leq q\} \) are compact and the following condition holds: \[ h\left( {x,{\partial }_{P}Q\left( x\right) }\right) \leq - W\left( x\right) \forall x \in \operatorname{comp}S. \] (a) Show that if such a pair \( \left( {Q, W}\right) \) exists, then another pair \( \left( {Q,\widetilde{W}}\right) \) exists for which \( \widetilde{W} \) is globally defined and locally Lipschitz on \( {\mathbb{R}}^{n} \), is equal to zero precisely on \( S \), and satisfies a linear growth condition. (b) Show that if a Lyapounov pair \( \left( {Q, W}\right) \) exists for \( S \), then any \( \alpha \in \) \( \operatorname{dom}Q \) can be steered at least asymptotically to \( S \) (i.e., there is a trajectory \( x\left( \cdot \right) \) which begins at \( \alpha \) and either reaches \( S \) in finite time, or else satisfies \( \left. {\mathop{\lim }\limits_{{t \rightarrow \infty }}{d}_{S}\left( {x\left( t\right) }\right) = 0}\right) \) . (c) A partial converse to (b) is available. Suppose that \( S \) has the property that any \( \alpha \in {\mathbb{R}}^{n} \) can be steered at least asymptotically to \( S \) . Let \( \varepsilon > 0 \) . Then the set \( S + \varepsilon \bar{B} \) admits a Lyapounov pair \( \left( {Q, W}\right) \) for which \( Q \) is finite. (Hint. Let \[ Q\left( \alpha \right) \mathrel{\text{:=}} \min \{ T \geq 0 : \text{ some trajectory }x\text{ has }x\left( 0\right) = \alpha, x\left( T\right) \in S + \varepsilon \bar{B}\} .) \] 11.16. Let \( \left( {\varphi, F}\right) \) be weakly decreasing, where \( \varphi : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) is continuous. Prove that given \( \alpha \in {\mathbb{R}}^{n} \), there is a trajectory \( x \) of \( F \) with \( x\left( 0\right) = \alpha \) such that the function \( t \mapsto \varphi \left( {x\left( t\right) }\right) \) is decreasing. (It is not known whether this is true when \( \varphi \) is merely assumed to belong to \( \mathcal{F}\left( {\mathbb{R}}^{n}\right) \) .) 11.17. For \( n = 1 \), let \( f\left( x\right) = 0 \) if \( x \in \left( {0,1}\right) \) and \( x \) is rational, \( f\left( x\right) = - 1 \) otherwise. Show that the following function \( \bar{x} \) is an Euler solution of the initial-value problem \( \dot{x} = f\left( x\right), x\left( 0\right) = 1 \) : \[ \bar{x}\left( t\right) = \left\{ \begin{array}{ll} 1 - t & \text{ if }0 \leq t \leq 1 \\ 0 & \text{ if }t \geq 1 \end{array}\right. \] Nonetheless, \( \bar{x} \) restricted to \( \lbrack 1,\infty ) \) is not an Euler solution of the initial-value problem \( \dot{x} = f\left( x\right), x\left( 1\right) = 0 \) . (The moral is that Euler solutions cannot be truncated to give Euler solutions.) 11.18. Let \( \varphi \) be continuous and let \( F \) be locally Lipschitz. Show by examples that the following monotonicity properties of the system \( \left( {\varphi, F}\right) \) are all distinct, and then go on to prove that each is characterized by its corresponding proximal Hamiltonian inequality holding for all \( x \) : (a) weakly decreasing: \( h\left( {x,{\partial }_{P}\varphi \left( x\right) }\right) \leq 0 \) ; (b) weakly increasing: \( H\left( {x,{\partial }^{P}\varphi \left( x\right) }\right) \geq 0 \) ; (c) strongly decreasing: \( H\left( {x,{\partial }_{P}\varphi \left( x\right) }\right) \leq 0 \) or \( H\left( {x,{\partial }^{P}\varphi \left( x\right) }\right) \leq 0 \) ; (d) strongly increasing: \( h\left( {x,{\partial }_{P}\varphi \left( x\right) }\right) \geq 0 \) or \( h\left( {x,{\partial }^{P}\varphi \left( x\right) }\right) \geq 0 \) ; (e) weakly predecreasing (that is, given any \( {x}_{0} \), there exists a trajectory \( x \) of \( F \) on \( ( - \infty ,0\rbrack \) such that \( x\left( 0\right) = {x}_{0} \) and \( \varphi \left( {x\left( t\right) }\right) \geq \varphi \left( {x}_{0}\right), t \leq 0) \) : \( h\left( {x,{\partial }^{P}\varphi \left( x\right) }\right) \leq 0 \) ; and (f) weakly preincreasing (the definition is analogous to (e)): \( H\left( {x,{\partial }_{P}\varphi \left( x\right) }\right) \geq 0 \) . In which of these cases is the Lipschitz hypothesis on \( F \) actually required? (Hint. Exercise 2.11 (c), (d) is relevant to (e) above.) 11.19. For a given initial time \( {t}_{0} \) and compact subset \( A \) of \( {\mathbb{R}}^{n} \), the reachable set \( \mathcal{R} \) from \( A \) is defined as follows: \[ \mathcal{R} \mathrel{\text{:=}} \left\{ {\left( {t, x\left( t\right) }\right) : t \geq {t}_{0}, x\text{ is a trajectory for }F\text{ on }\left\lbrack {{t}_{0}, t}\right\rbrack, x\left( {t}_{0}\right) \in A}\right\} . \] Here we consider \( F \) depending on both \( t \) and \( x \) . We denote the "slice" at time \( T
1012_(GTM178)Nonsmooth Analysis and Control Theory
90
) such that \( x\left( 0\right) = {x}_{0} \) and \( \varphi \left( {x\left( t\right) }\right) \geq \varphi \left( {x}_{0}\right), t \leq 0) \) : \( h\left( {x,{\partial }^{P}\varphi \left( x\right) }\right) \leq 0 \) ; and (f) weakly preincreasing (the definition is analogous to (e)): \( H\left( {x,{\partial }_{P}\varphi \left( x\right) }\right) \geq 0 \) . In which of these cases is the Lipschitz hypothesis on \( F \) actually required? (Hint. Exercise 2.11 (c), (d) is relevant to (e) above.) 11.19. For a given initial time \( {t}_{0} \) and compact subset \( A \) of \( {\mathbb{R}}^{n} \), the reachable set \( \mathcal{R} \) from \( A \) is defined as follows: \[ \mathcal{R} \mathrel{\text{:=}} \left\{ {\left( {t, x\left( t\right) }\right) : t \geq {t}_{0}, x\text{ is a trajectory for }F\text{ on }\left\lbrack {{t}_{0}, t}\right\rbrack, x\left( {t}_{0}\right) \in A}\right\} . \] Here we consider \( F \) depending on both \( t \) and \( x \) . We denote the "slice" at time \( T \) of \( \mathcal{R} \) by \( {\mathcal{R}}_{T} \) ; that is, \[ {\mathcal{R}}_{T} \mathrel{\text{:=}} \{ x : \left( {T, x}\right) \in \mathcal{R}\} \] Then clearly \[ {\mathcal{R}}_{T} = \bigcup \left\{ {\mathcal{A}\left( {{t}_{0},{x}_{0};T}\right) : {x}_{0} \in A}\right\} \] We will see that the intrinsic invariance properties of \( \mathcal{R} \) allow us to characterize it via a certain Hamilton-Jacobi relationship satisfied by its proximal normal vectors. (a) Prove that \( \mathcal{R} \) is closed and that each \( {\mathcal{R}}_{T} \) is compact. Show that \( {\mathcal{R}}_{{t}_{0}} = A \) . (b) Prove that \( \mathcal{R} \) is uniformly bounded near \( {t}_{0} \) ; i.e., there exists \( \varepsilon > 0 \) and a compact set \( C \) such that \[ {\mathcal{R}}_{T} \subseteq C\forall T \in \left\lbrack {{t}_{0},{t}_{0} + \varepsilon }\right\rbrack \] (c) Let \( S \) be a closed subset of \( \left\lbrack {{t}_{0},\infty }\right) \times {\mathbb{R}}^{n} \) which is uniformly bounded near \( {t}_{0} \), and for which \( {S}_{{t}_{0}} = A \) . Prove that for any \( \varepsilon > 0 \) there exists \( \delta > 0 \) such that \( {S}_{T} \subseteq A + {\varepsilon B} \) for all \( T \in \left\lbrack {{t}_{0},{t}_{0} + \delta }\right\rbrack \) . (d) Prove the following: Theorem. Let \( F \) be locally Lipschitz. Then \( \mathcal{R} \) is the unique closed subset \( S \) of \( \left\lbrack {{t}_{0},\infty }\right) \times {\mathbb{R}}^{n} \) which is uniformly bounded near \( {t}_{0} \) and satisfies: (i) \( \theta + H\left( {t, x,\zeta }\right) = 0\forall \left( {\theta ,\zeta }\right) \in {N}_{S}^{P}\left( {t, x}\right) ,\forall \left( {t, x}\right) \in \left( {{t}_{0},\infty }\right) \times {\mathbb{R}}^{n} \) ; and (ii) \( {S}_{{t}_{0}} = A \) . 11.20. Let \( \left( {{Q}_{1},{W}_{1}}\right) \) and \( \left( {{Q}_{2},{W}_{2}}\right) \) be Lyapounov pairs for \( {x}^{ * } \) . Prove that \( \left( {\min \left\{ {{Q}_{1},{Q}_{2}}\right\} ,\min \left\{ {{W}_{1},{W}_{2}}\right\} }\right) \) is also a Lyapounov pair for \( {x}^{ * } \) . 11.21. For \( n = 2, F \) induces the lower Hamiltonian \[ h\left( {x, y, p, q}\right) = - \left| {\left( {x - y}\right) \left( {p - 1}\right) }\right| . \] (a) Find \( F \) . (b) Let \[ {\varphi }_{1}\left( {\tau ,\alpha ,\beta }\right) \mathrel{\text{:=}} {e}^{{2\tau } - 2}\left| {\alpha - \beta }\right| ,\;{\varphi }_{2}\left( {\tau ,\alpha ,\beta }\right) \mathrel{\text{:=}} - {e}^{-{2\tau } + 2}\left| {\alpha - \beta }\right| . \] Which of \( {\varphi }_{1},{\varphi }_{2} \) is the value function of a problem of the form \[ \text{minimize}\{ \ell \left( {x\left( 1\right) }\right) : \dot{x} \in F\left( x\right), x\left( \tau \right) = \left( {\alpha ,\beta }\right) \} \text{?} \] 11.22. In the context of Theorem 9.1, suppose that \( F\left( x\right) \) has the form \( f\left( {x, U}\right) \), where \( U \) is a compact subset of \( {\mathbb{R}}^{m} \) and \( f \) is continuously differentiable. Suppose in addition that \( F\left( x\right) \) is strictly convex for each \( x \), in the following sense: \[ v, w \in F\left( x\right) ,\;v \neq w,\;\lambda \in \left( {0,1}\right) \Rightarrow {\lambda v} + \left( {1 - \lambda }\right) w \in \operatorname{int}F\left( x\right) . \] Derive from the Hamiltonian inclusion of Theorem 9.1 the following conclusion: there exists a measurable function \( u\left( \cdot \right) \) on \( \left\lbrack {a, b}\right\rbrack \) with values in \( U \) such that for almost every \( t \) in \( \left\lbrack {a, b}\right\rbrack \) we have \[ \dot{x}\left( t\right) = f\left( {x\left( t\right), u\left( t\right) }\right) ,\; - \dot{p}\left( t\right) = {f}_{x}^{\prime }{\left( x\left( t\right), u\left( t\right) \right) }^{ * }p\left( t\right) , \] \[ \mathop{\max }\limits_{{{u}^{\prime } \in U}}\left\langle {p\left( t\right), f\left( {x\left( t\right) ,{u}^{\prime }}\right) }\right\rangle = \langle p\left( t\right) ,\dot{x}\left( t\right) \rangle . \] These conclusions (together with the transversality condition) constitute the Pontryagin Maximum Principle; they can be obtained under weaker assumptions than those in force here. (Hint. Problem 2.9.13.) 11.23. Consider the problem of minimizing \[ \ell \left( {x\left( b\right) }\right) + {\int }_{a}^{b}\varphi \left( {x\left( t\right) }\right) {dt} \] over those trajectories of \( F \) on \( \left\lbrack {a, b}\right\rbrack \) satisfying \( x\left( a\right) = {x}_{0} \) . Here, \( {x}_{0} \) is given, and \( \ell \) and \( \varphi \) are given locally Lipschitz functions. Let \( x\left( \cdot \right) \) be a solution. Use the device introduced in the proof of Theorem 10.3 (absorbing \( \varphi \) into the dynamics) to prove the following necessary conditions: there exists an arc \( p \) such that \[ \left( {-\dot{p}\left( t\right) ,\dot{x}\left( t\right) }\right) \in {\partial }_{C}H\left( {x\left( t\right), p\left( t\right) }\right) - {\partial }_{C}\varphi \left( {x\left( t\right) }\right) \times \{ 0\} ,\;t \in \left\lbrack {a, b}\right\rbrack \text{ a.e.,} \] \[ H\left( {x\left( t\right), p\left( t\right) }\right) - \varphi \left( {x\left( t\right) }\right) = \text{ constant,}\;t \in \left\lbrack {a, b}\right\rbrack , \] \[ - p\left( b\right) \in {\partial }_{L}\ell \left( {x\left( b\right) }\right) . \] (Hint. Without loss of generality, we can assume that \( \varphi \) satisfies a linear growth condition.) 11.24. With \( n = 1 \), let \( x\left( \cdot \right) \) solve the problem of minimizing \[ - {\beta x}\left( 1\right) + {\int }_{0}^{1}\left| {x\left( t\right) }\right| {dt} \] over those arcs \( x \) on \( \left\lbrack {0,1}\right\rbrack \) satisfying \( x\left( 0\right) = \alpha \) and \( \left| {\dot{x}\left( t\right) }\right| \leq 1, t \in \left\lbrack {0,1}\right\rbrack \) a.e., where \( \alpha \) and \( \beta \) are given constants. (a) Prove that \( x \) is piecewise affine, and that only three certain values are possible for the slopes of the affine portions. (Hint. Study the curves \( H\left( {x, p}\right) - \left| x\right| = \) constant.) (b) Find the unique solution \( x \) when \( \alpha \) and \( \beta \) are positive. 11.25. Let \( F \) be locally Lipschitz, and assume that the system is globally null controllable. Then the minimal time function \( Q \) (see Exercise 10.7) is everywhere finite. Prove that \( Q \) is the unique function \( \varphi : {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) satisfying the following conditions: (i) \( \varphi \in \mathcal{F}\left( {\mathbb{R}}^{n}\right) ,\varphi \left( 0\right) = 0,\varphi \left( x\right) > 0 \) for \( x \neq 0 \) ; (ii) \( \mathop{\liminf }\limits_{\substack{{{x}^{\prime } \rightarrow x} \\ {{x}^{\prime } \neq 0} }}\varphi \left( {x}^{\prime }\right) = 0 \) ; and (iii) \( h\left( {x,{\partial }_{P}\varphi \left( x\right) }\right) = - 1 \) on \( {\mathbb{R}}^{n} \smallsetminus \{ 0\} \) . 11.26. We will prove that when the origin is normal, the minimal-time function is continuous at 0 . The context is that of Theorem 10.3, in which it was shown that for certain \( b > 0 \), the value function \( {V}_{b} \) used in the proof is Lipschitz near 0 : for some \( {\delta }_{b} > 0 \), we have \[ 0 \leq {V}_{b}\left( \alpha \right) \leq {K}_{b}\parallel \alpha \parallel \;\text{ when }\parallel \alpha \parallel \leq {\delta }_{b} \] (a) Prove that this conclusion holds for all \( b > 0 \) sufficiently small. (b) If \( Q \) is the minimal-time function, then show that \[ Q\left( \alpha \right) \leq \sqrt{{K}_{b}\parallel \alpha \parallel } + b\;\text{ when }\parallel \alpha \parallel < {\delta }_{b}. \] (c) Given \( \varepsilon > 0 \), pick \( b > 0 \) so that \( \sqrt{b} + b < \varepsilon \) . Then, for \( \parallel \alpha \parallel < \) \( \min \left( {{\delta }_{b}, b/{K}_{b}}\right) \), we have \( Q\left( \alpha \right) < \varepsilon \) . Deduce that \( Q \) is continuous at 0 . 11.27. Let \( {\mathcal{R}}_{0} \) be the reachable set from the origin: \[ {\mathcal{R}}_{0} \mathrel{\text{:=}} \{ \left( {t, x\left( t\right) }\right) : t \geq 0, x\text{ is a trajectory,}x\left( 0\right) = 0\} . \] We assume that \( F \) is locally Lipschitz and autonomous. Suppose that \( \left( {T,\beta }\right) \) is a boundary point of \( {\mathcal{R}}_{0}, T > 0 \) . (a) Prove the existence of an arc \( p \) on \( \left\lbrack {0, T}\right\rbrack \), and a trajectory \( x \) of \( F \) on \( \left\lbrack {0, T}\right\rbrack \) with \( x\left( 0\right) = 0, x\left( T\right) = \beta \) such that for some constant \( h \) we have \( \left| h\right| + \parallel p{\parallel }_{\infty } \neq 0 \) and \[ \left( {-\dot{p}\left( t\right) ,\dot{x}\left( t\right) }\right) \in {\partial }_{C}H\left( {x\left( t\right), p\left( t\right) }\right) ,\;t \in \left\lbrack {0, T}\right\rbrack \text{ a.e.,} \] \[ H\left( {x\left( t\right), p\left( t\right) }\right) = h,\;t \in \left\lbrack {0, T}\right\rbrack , \] \[ \left( {h, - p\left( T\right) }\
1012_(GTM178)Nonsmooth Analysis and Control Theory
91
m the origin: \[ {\mathcal{R}}_{0} \mathrel{\text{:=}} \{ \left( {t, x\left( t\right) }\right) : t \geq 0, x\text{ is a trajectory,}x\left( 0\right) = 0\} . \] We assume that \( F \) is locally Lipschitz and autonomous. Suppose that \( \left( {T,\beta }\right) \) is a boundary point of \( {\mathcal{R}}_{0}, T > 0 \) . (a) Prove the existence of an arc \( p \) on \( \left\lbrack {0, T}\right\rbrack \), and a trajectory \( x \) of \( F \) on \( \left\lbrack {0, T}\right\rbrack \) with \( x\left( 0\right) = 0, x\left( T\right) = \beta \) such that for some constant \( h \) we have \( \left| h\right| + \parallel p{\parallel }_{\infty } \neq 0 \) and \[ \left( {-\dot{p}\left( t\right) ,\dot{x}\left( t\right) }\right) \in {\partial }_{C}H\left( {x\left( t\right), p\left( t\right) }\right) ,\;t \in \left\lbrack {0, T}\right\rbrack \text{ a.e.,} \] \[ H\left( {x\left( t\right), p\left( t\right) }\right) = h,\;t \in \left\lbrack {0, T}\right\rbrack , \] \[ \left( {h, - p\left( T\right) }\right) \in {N}_{{\mathcal{R}}_{0}}^{L}\left( {T,\beta }\right) . \] (Hint. Consider first the case in which \( {N}_{{\mathcal{R}}_{0}}^{P}\left( {T,\beta }\right) \) is nontrivial; then some point \( \left( {\tau ,\gamma }\right) \in \operatorname{comp}{\mathcal{R}}_{0} \) admits \( \left( {T,\beta }\right) \) as the closest point in \( {\mathcal{R}}_{0} \) ; apply Exercise 9.16.) (b) Deduce that some solution \( x \) of the minimal-time problem (Exercise 10.7) admits an arc \( p \) and a constant \( h \) with \( \left| h\right| + \parallel p{\parallel }_{\infty } \neq 0 \) such that \[ \left( {-\dot{p}\left( t\right) ,\dot{x}\left( t\right) }\right) \in {\partial }_{C}H\left( {x\left( t\right), p\left( t\right) }\right) ,\;t \in \left\lbrack {0, T}\right\rbrack \text{ a.e.,} \] \[ H\left( {x\left( t\right), p\left( t\right) }\right) = h,\;t \in \left\lbrack {0, T}\right\rbrack . \] (c) Consider the problem of finding the function \( x : \left\lbrack {0, T}\right\rbrack \rightarrow \mathbb{R} \) which steers an initial position \( \alpha = x\left( 0\right) \) and an initial velocity \( v = \dot{x}\left( 0\right) \) to the origin and at rest in minimal time \( T \) under the constraint \( \left| {\ddot{x}\left( t\right) }\right| \leq 1 \) . Interpret it as a special case of the minimal-time problem, with \( n = 2 \) and \( F\left( {x, y}\right) \mathrel{\text{:=}} \{ \left( {y, u}\right) : \left| u\right| \leq 1\} \) . (d) Interpret the necessary conditions of part (b) for this case. (Hint. Exercise 9.9.) Proceed to find the unique solution of the problem, as a function of the initial condition \( \left( {\alpha, v}\right) \) . (e) Show that the origin is normal. (f) By calculating the minimal-time function explicitly, verify that it is continuous, but not Lipschitz, at zero. 11.28. Let \( f : {\mathbb{R}}^{m} \times {\mathbb{R}}^{n} \rightarrow \mathbb{R} \) be locally Lipschitz, and suppose that \( f\left( {x, y}\right) \) is concave as a function of \( x \) and convex as a function of \( y \) . Prove that \[ {\partial }_{C}f\left( {x, y}\right) = {\partial }_{C}f\left( {\cdot, y}\right) \left( x\right) \times {\partial }_{C}f\left( {x, \cdot }\right) \left( y\right) . \] (Hint. See Problem 2.9.15, as well as the proof of Lemma 9.5.) 11.29. We consider the context of Theorem 9.2, under the additional assumption that the set \( G \mathrel{\text{:=}} \operatorname{graph}F \) is convex. (a) Prove that \( H\left( {x, p}\right) \) is concave as a function of \( x \) . (b) Show that the Hamiltonian inclusion is equivalent to the relations \[ - \dot{p}\left( t\right) \in {\partial }_{C}H\left( {\cdot, p\left( t\right) }\right) \left( {x\left( t\right) }\right) ,\;\dot{x}\left( t\right) \in {\partial }_{C}H\left( {x\left( t\right) , \cdot }\right) \left( {p\left( t\right) }\right) . \] (c) Show that the Hamiltonian inclusion is also equivalent to \[ \left( {\dot{p}\left( t\right), p\left( t\right) }\right) \in {N}_{G}^{P}\left( {x\left( t\right) ,\dot{x}\left( t\right) }\right) . \] (d) If in addition it is assumed that \( \ell \left( \cdot \right) \) is convex, then prove that any arc \( x \) admissible for (P) which satisfies (together with some arc \( p \) ) the Hamiltonian inclusion and the transversality condition is a solution of (P) (i.e., the necessary conditions of Theorem 9.1 are also sufficient for optimality). 11.30. Let \( V \) be the value function of \( §7 \) : \[ V\left( {\tau ,\alpha }\right) \mathrel{\text{:=}} \min \{ \ell \left( {x\left( T\right) }\right) : \dot{x} \in F\left( x\right), x\left( \tau \right) = \alpha \} , \] where \( \ell \) and \( F \) are locally Lipschitz. Let \( M\left( {\tau ,\alpha }\right) \) denote the set of arcs \( x \) on \( \left\lbrack {\tau, T}\right\rbrack \) satisfying \( x\left( \tau \right) = \alpha \) and admitting an arc \( p \) such that \[ \left( {-\dot{p}\left( t\right) ,\dot{x}\left( t\right) }\right) \in {\partial }_{C}H\left( {x\left( t\right), p\left( t\right) }\right) ,\;t \in \left\lbrack {\tau, T}\right\rbrack , \] \[ H\left( {x\left( t\right), p\left( t\right) }\right) = \text{ constant,}\;t \in \left\lbrack {\tau, T}\right\rbrack , \] \[ - p\left( T\right) \in {\partial }_{L}\ell \left( {x\left( T\right) }\right) . \] Prove that \[ V\left( {\tau ,\alpha }\right) = \min \{ \ell \left( {x\left( T\right) }\right) : x \in M\left( {\tau ,\alpha }\right) \} . \] (The fact that \( V \), the solution of the Hamilton-Jacobi equation, is generated by solutions of a Hamiltonian system is known in a classical setting as the method of characteristics.) 11.31. In the context of Theorem 5.5, suppose that \( Q \) and \( W \) are continuously differentiable. Prove that the system \( \left( {F,{x}^{ * }}\right) \) admits a globally stabilizing feedback \( \widehat{v}\left( \cdot \right) \) ; i.e., a selection \( \widehat{v} \) for \( F \) all of whose Euler solutions on \( \lbrack 0,\infty ) \), for any initial condition, tend to \( {x}^{ * } \) as \( t \rightarrow \infty \) . (Hint. Consider the multifunction \[ \widehat{F}\left( x\right) \mathrel{\text{:=}} \{ v \in F\left( x\right) : \langle \nabla Q\left( x\right), v\rangle \leq - W\left( x\right) \} .) \] We remark that a "nice" system which is globally asymptotically controllable to the origin need not admit a smooth Lyapounov pair. An example is the non-holonomic integrator, in which \( n = 3 \) and \[ F\left( x\right) \mathrel{\text{:=}} \left\{ {\left( {{u}_{1},{u}_{2},{x}_{1}{u}_{2} - {u}_{1}{x}_{2}}\right) : \begin{Vmatrix}\left( {{u}_{1},{u}_{2}}\right) \end{Vmatrix} \leq 1}\right\} . \] Notes and Comments To be neglected before one's time, must be very vexatious. The theory of generalized gradients, proposed by Clarke in 1973, demonstrated the possibility of developing a useful calculus in a fully nonsmooth, nonconvex setting. This theory has served as the pattern for subsequent ones, by indicating how and why to treat functions and sets on an equal footing, how to pass back and forth from one to the other, and in so doing obtain a coherent, complete calculus. Certain elements of this design were foreshadowed by convex analysis, in the work of Moreau and Rockafellar. But the opportunity to shed convexity hypotheses greatly expanded the potential applications of the subject. The basic construct of proximal analysis, the proximal normal vector, also appears in Clarke's early work, under the name "perpendicular." It was used there to generate the normal cone (and hence, via epigraphs, subgra-dients). In that sense, proximal subgradients were implicit from the beginning. However, the development of proximal calculus per se came later, and it gradually became clear that proximal analysis should be viewed as a distinct branch of the subject in its own right. The work of Rockafellar, Ioffe, Mordukhovich, and Borwein is to be mentioned in this regard. We will not attempt to list, in an introductory text, the hundreds of references, nor the dozens of other names, that could justifiably be mentioned. However, the Bibliography contains a selection of recent articles and books on both theory and applications that can serve as a starting point for the interested reader. In the notes below, when no citation is given, the result (or a variant of it) is likely to be found in one or both of the two basic sources for this work, [C4] and [C5], which also contain many other references and related topics. Other general works that we recommend to the reader include, for non-smooth and nonlinear analysis, Aubin [A2] (which emphasizes games and economics), Demyanov and Rubinov [DR] and Loewen [L1] (emphasis on optimal control), and Mäkelä and Neittaanmäki [MN] (emphasizes applications). Convex analysis is treated, for example, by Hiriart-Urruty and Lemaréchal in [HUL]. Phelps [Ph] is an excellent source for differentiability. As standard references on control theory, we suggest Fleming and Soner [FS], Roxin [Ro], Sontag [So], and Zabczyk [Z]. The books [HL], [PBGM], and [Y] are old favorites of ours. ## Chapter 0 The elementary discussion of Dini derivates in Boas [Bo] is a nice one; the definitive references in this area remain the classic books of Saks [Sa] and Bruckner [Bru]. The argument used to prove Theorem 1.1 is motivated by that in [CR]. The existence theorem in the calculus of variations that is alluded to appears in [C6]. The tangential characterization of flow-invariance given by Theorem 2.3 has a history of rediscovery, beginning with Nagumo in 1942. The normality characterization of Theorem 2.4 goes back to Bony. Optimization of eigenvalues is discussed in [BO], [HY], and [O]; see Cox and Overton [CO] and [Cx] for the strongest column. Torricelli's table was inspired by an example in Lyusternik [Ly]. Tikhomirov's book [T1] is a charming, elementary introduction to optimization. ## Chapter 1 The proof technique of Theorem 3.1 is now a standard one (see [LS]) in the subject of minimization prin
1012_(GTM178)Nonsmooth Analysis and Control Theory
92
Zabczyk [Z]. The books [HL], [PBGM], and [Y] are old favorites of ours. ## Chapter 0 The elementary discussion of Dini derivates in Boas [Bo] is a nice one; the definitive references in this area remain the classic books of Saks [Sa] and Bruckner [Bru]. The argument used to prove Theorem 1.1 is motivated by that in [CR]. The existence theorem in the calculus of variations that is alluded to appears in [C6]. The tangential characterization of flow-invariance given by Theorem 2.3 has a history of rediscovery, beginning with Nagumo in 1942. The normality characterization of Theorem 2.4 goes back to Bony. Optimization of eigenvalues is discussed in [BO], [HY], and [O]; see Cox and Overton [CO] and [Cx] for the strongest column. Torricelli's table was inspired by an example in Lyusternik [Ly]. Tikhomirov's book [T1] is a charming, elementary introduction to optimization. ## Chapter 1 The proof technique of Theorem 3.1 is now a standard one (see [LS]) in the subject of minimization principles, for which a general reference is [DGZ]. The best known such principle is Ekeland's [E]; however, it is of limited interest in proximal analysis since it introduces a term with a "corner," in contrast to the "smooth variational principle" of Borwein and Preiss [BP]. We follow [CLW] in deriving this result from proximal density. This article contains an example of a \( {C}^{1} \) function on the line whose proximal subdiffer-ential is empty except on a set which is small, both in the sense of measure and of category. Theorem 7.3 has many predecessors; the approach used here appears in [CSW], [RW], and [W]. The result also follows immediately from the Mean Value Inequality of §3.2. The fuzzy Sum Rule Theorem 8.3 goes back to Ioffe in its essentials, while the limiting Sum Rule 10.1 is due to Mordukhovich. Problem 11.23 is a result due to Rockafellar in finite dimensions. ## Chapter 2 The results of this chapter follow closely those in Clarke [C4], which has significantly more material, however. The Chain Rule 2.5 is an improvement on the original treatment. Because the notation \( \partial ,{N}_{S} \), and \( {T}_{S} \) is now so widely used, we have opted to retain it, while simultaneously introducing the alternate notation \( {\partial }_{C},{N}_{S}^{C} \), and \( {T}_{S}^{C} \) for cases in which other constructs also intervene (as in Chapter 4). ## Chapter 3 Constrained optimization is treated at greater length in Chapter 6 of [C4]. The Mean Value Inequality of Theorem 2.6 is due to Clarke and Ledyaev [CL2]. Surprisingly perhaps, the case in which the values at two different sets are compared appears to be quite different; it is treated in [CL1]. The results of \( §3 \) have innumerable antecedents, though the specific approach seems to be new. The Lipschitz Inverse Function Theorem is due to Clarke. For Theorem 4.2, see, for example, Subbotin [Su]. Theorem 4.16, and the resulting proof of Rademacher's Theorem, are drawn from Bessis and Clarke [BC]. Section 5 follows [C4] for the most part. We refer to [C5] for further results in the calculus of variations. It was Rockafellar who first pointed out the importance of the tangent cone \( {T}_{S}^{C} \) having a nonempty interior, and the properties that result from this. What we have called "wedged" is a new term that we are proposing here instead of "epi-Lipschitz." Theorem 6.12 is due to Aubin and Clarke [ACl], and the function in Problem 7.30 is due to Rockafellar. ## Chapter 4 A great deal of work on what we have called weak invariance has been done under the label viability, see Aubin [A1] and Aubin and Cellina [AC]. The book by Deimling [D] is another valuable reference for differential inclusions (note: neither "invariance" nor "viability" is employed here; the author simply speaks of existence). The concept of Euler solutions is borrowed from Krasovskii and Subbotin [KS], where it plays an important role in differential games; our proximal aiming is inspired by their "extremal aiming method." Theorem 2.10 is an amalgam of several results, including certain ones for the case of a function (see the comments for Chapter 0), as well as theorems of Haddad and of Veliov [V] (to whom is first due the proximal criterion). See [CLSW] for further comment. The Strong Invariance Theorem 3.8 has its antecedents in [C2]. The results of \( §4 \) are taken from [CLS1]. We remark that all the results of this chapter can be extended to the case in which the linear growth hypothesis is dropped, by taking account of possible finite "blow-up" times. Several schools have participated over the past decades in developing the nonsmooth theory of the Hamilton-Jacobi equation. The state of the art circa 1976 is described in [B] where the "almost everywhere" type of (Lipschitz) solution dominates, and where Fleming's "artificial viscosity" approximation method is outlined. To our knowledge, the first truly pointwise (subdifferential) definition of the generalized solution (proposed by Clarke, again for Lipschitz functions) appeared in 1977 [H] in connection with the verification method. These turn out to be semisolutions in the viscosity sense. The approach is synthesized by Clarke and Vinter in [CV]. In 1980 Subbotin (see [Su]) inaugurated the two-sided approach to defining a nonsmooth (Lipschitz) solution, with Dini derivates and in the context of differential games. Subsequently, Crandall and Lions (see [CIL]) developed further Fleming's approach; the (eventual) two-sided, subdifferential approach to defining viscosity solutions, carried out for continuous functions and with the attendant uniqueness theorem, constituted a breakthrough that vindicated the nonsmooth analysis approach to the issue. It appears to be Barron and Jensen [BJ] who first demonstrated the (surprising) possibility of giving a single subdifferential characterization in certain cases, and for merely lower semicontinuous solutions. The results of \( §7 \) owe an intellectual debt to them, and also to Subbotin, who first stressed the relevance of invariance. See also [AF2], [Ba], [BaS], [CaF], [CD], and [He] for other examples of current research directions. Theorem 9.1 is due to Clarke; Theorem 10.3 is taken from [CL]. A number of results relating the adjoint variable to the sensitivity of the problem are due to Clarke and Vinter. We refer to [C5] for these and other refinements of the necessary conditions in optimal control, and for examples of their use, under less stringent hypotheses; in this connection, see [C3] for the maximum principle. In particular, it is not assumed that the velocity set is convex. To put this another way, the current chapter deals with the "relaxed problem" only. The forthcoming book of Vinter [Vi] surveys the question of necessary conditions in optimal control, including results of Ioffe, Loewen, Mordukhovich, and Rockafellar aimed at refining the Hamiltonian inclusion. It is a celebrated theorem of Brockett that (smooth) systems which are globally asymptotically controllable to the origin need not admit a continuous feedback control which stabilizes the system. In [CLSS], it is shown that time-independent stabilizing feedbacks do exist, if they are allowed to be discontinuous and if the corresponding solution concept is defined as in Chapter 4. This is achieved essentially as follows: a result of Sontag is invoked to deduce the existence of a continuous Lyapounov pair, and then an analogue of Problem 11.31 is carried out. Insight on the existence or otherwise of smooth Lyapounov pairs is provided in [CLS2]. See also [Ber], [CPT], [Kry], [KS], and [RolV] for other approaches to feedback construction. List of Notation \( {\operatorname{proj}}_{S}\left( x\right) \) \( {d}_{S}\left( x\right) \) or \( d\left( {x;S}\right) \) \( {N}_{S}^{P}\left( x\right) \) or \( {N}^{P}\left( {x;S}\right) \) \( {\partial }_{P}f\left( x\right) \) \( {\partial }^{P}f\left( x\right) \) \( \operatorname{dom}f \) \( \operatorname{gr}f \) \( \operatorname{epi}f \) \( {f}^{ \circ }\left( {x;v}\right) \) \( {h}_{S}\left( \cdot \right) ,{H}_{S}\left( \cdot \right) \) \( {N}_{S}\left( x\right) \) or \( {N}_{S}^{C}\left( x\right) \) We list here the principal constructs that appear in the book. projection of \( x \) onto \( S \) . distance from \( x \) to \( S \) . proximal normal cone to \( S \) at \( x \) . proximal subdifferential of \( f \) at \( x \) . proximal superdifferential of \( f \) at \( x \) . (effective) domain of \( f \) . graph of \( f \) . epigraph of \( f \) . \( \mathcal{F}\left( U\right) \; \) all \( f : U \rightarrow ( - \infty ,\infty \rbrack \) that are lower semicontinuous and not identically \( + \infty \) . limiting subdifferential of \( f \) at \( x \) . \( {N}_{S}^{L}\left( x\right) \; \) limiting normal cone to \( S \) at \( x \) \( {f}^{\prime }\left( {x;v}\right) \; \) directional derivative of \( f \) at \( x \) in direction \( v \) . \( {f}_{G}^{\prime }\left( x\right) \; \) Gâteaux derivative of \( f \) at \( x \) . \( {f}^{\prime }\left( x\right) \; \) Fréchet derivative of \( f \) at \( x \) . \( {I}_{S}\left( \cdot \right) \) or \( I\left( {\cdot ;S}\right) \; \) indicator function of a set \( S \) . generalized directional derivative of \( f \) at \( x \) in direction \( v \) . lower and upper support functions of a set \( S \) . \( \partial f\left( x\right) \) or \( {\partial }_{C}f\left( x\right) \; \) generalized gradient of \( f \) at \( x \) . \( {T}_{S}\left( x\right) \) or \( {T}_{S}^{C}\left( x\right) \; \) (generalized) tangent cone to \( S \) at \( x \) . \( {T}_{S}^{B}\left( x\right) \; \) Bouligand tangent cone to \( S \) at \( x \) . (generalized) normal cone to \( S \) at \( x \) . 264 List of Notation \( {Df}\left( {x;v}\right) \) subderivate of \( f \) at \( x \) in direction \( v \) . \( {\partial }_{D}f\left( x\right) \) directional subdifferential of \( f \) at \( x \) . \( \nabla f\left( x\right) \; \) gradient ve
1012_(GTM178)Nonsmooth Analysis and Control Theory
93
; \) Gâteaux derivative of \( f \) at \( x \) . \( {f}^{\prime }\left( x\right) \; \) Fréchet derivative of \( f \) at \( x \) . \( {I}_{S}\left( \cdot \right) \) or \( I\left( {\cdot ;S}\right) \; \) indicator function of a set \( S \) . generalized directional derivative of \( f \) at \( x \) in direction \( v \) . lower and upper support functions of a set \( S \) . \( \partial f\left( x\right) \) or \( {\partial }_{C}f\left( x\right) \; \) generalized gradient of \( f \) at \( x \) . \( {T}_{S}\left( x\right) \) or \( {T}_{S}^{C}\left( x\right) \; \) (generalized) tangent cone to \( S \) at \( x \) . \( {T}_{S}^{B}\left( x\right) \; \) Bouligand tangent cone to \( S \) at \( x \) . (generalized) normal cone to \( S \) at \( x \) . 264 List of Notation \( {Df}\left( {x;v}\right) \) subderivate of \( f \) at \( x \) in direction \( v \) . \( {\partial }_{D}f\left( x\right) \) directional subdifferential of \( f \) at \( x \) . \( \nabla f\left( x\right) \; \) gradient vector of \( f \) at \( x \) . \( h, H\; \) lower and upper Hamiltonians. We remark that at least four people in the field agree on these notational choices, which is certainly a record. ## Bibliography Artstein, Z., Stabilization with relaxed controls, Nonlinear Anal. 7 (1983), 1163–1173. Attouch, H., Viscosity solutions of minimization problems, SIAM J. Optim. 6 (1996), 769-806. [A1] Aubin, J.-P., Viability Theory, Birkhäuser, Boston, 1991. 2] Aubin, J. P., Optima and Equilibria-An Introduction to Nonlinear Analysis, Graduate Texts in Mathematics, vol. 140, Springer-Verlag, New York, 1993. y] Aubin, J.-P. and Cellina, A., Differential Inclusions, Springer-Verlag, Berlin, 1994. \( \left\lbrack \mathrm{{ACl}}\right\rbrack \) Aubin, J.-P. and Clarke, F. H., Monotone invariant solutions to differential inclusions, J. London Math. Soc. 16 (2) (1977), 357- 366. [AF1] Aubin, J.-P. and Frankowska, H., Set-Valued Analysis, Birkhäuser, Boston, 1990. [AF2] Aubin, J.-P. and Frankowska, H., Partial differential inclusions governing feedback controls, J. Convex Anal. 2 (1995), 19-40. \( \left\lbrack \mathrm{{BaS}}\right\rbrack \) Bardi, M. and Soravia, P., A comparison result for Hamilton-Jacobi equations and applications to some differential games lacking controllability, Funkcial. Ekvac. 37 (1994), 19-43. 266 Bibliography [Ba] Barles, G., Discontinuous viscosity solutions of first-order Hamilton-Jacobi equations: A guided visit, Nonlinear Anal. 20 (1993), 1123-1134. [BJ] Barron, E. N. and Jensen, R., Optimal control and semicontinuous viscosity solutions, Proc. Amer. Math. Soc. 113 (1991), 397-402. [BB] Basar, T. and Bernhard, P., \( {H}^{\infty } \) -Optimal Control and Related Minimax Design Problems, Birkhäuser, Berlin, 1991. [B] Benton, S. H., The Hamilton-Jacobi Equation: A Global Approach, Academic Press, New York, 1977. [Ber] Berkovitz, L. D., Optimal feedback controls, SIAM J. Control Optim. 27 (1989), 991-1006. [BC] Bessis, D. and Clarke, F. H., Partial subdifferentials, derivates, and Rademacher's Theorem, Trans. Amer. Math. Soc., in press. [BEFB] Boyd, S., El Ghaoui, L., Feron, E. and Balakrishnan, V., Linear Matrix Inequalities in System and Control Theory, SIAM, Philadelphia, PA, 1994. [Bo] Boas, R. P., A Primer of Real Functions, Carus Mathematical Monographs, Mathematical Association of America, Washington, DC, 1960. [BoCo] Bonnisseau, J.-M. and Cornet, B., Existence of marginal cost pricing equilibria: The nonsmooth case, Internat. Econom. Rev. 31 (1990), 685-708. [BI] Borwein, J. M. and Ioffe, A., Proximal analysis in smooth spaces, Set-Valued Anal. 4 (1996), 1-24. [BP] Borwein, J. M. and Preiss, D., A smooth variational principle with applications to subdifferentiability and to differentiability of convex functions, Trans. Amer. Math. Soc. 303 (1987), 517-527. [BZ] Borwein, J. M. and Zhu, Q. J., Variational analysis in nonreflex-ive spaces and applications to control problems with \( {L}^{1} \) perturbations, Nonlinear Anal. 28 (1997), 889-915. [Br] Brézis, H., Analyse fonctionnelle: Théorie et applications, Masson, Paris, 1983. [Bru] Bruckner, A., Differentiation of Real Functions, 2nd ed., CRM Monograph Series 5, American Mathematical Society, Providence, RI, 1994. ) Burke, J. V. and Overton, M. L., Differential properties of the spectral abscissa and the spectral radius for analytic matrix-valued mappings, Nonlinear Anal. 23 (1994), 467-488. [By] Byrnes, C. I., On the control of certain infinite dimensional systems by algebro-geometric techniques, Amer. J. Math. 100 (1978), 1333-1381. \( \left\lbrack \mathrm{{CD}}\right\rbrack \) Cannarsa, P. and DaPrato, G., Second-order Hamilton-Jacobi equations in infinite dimensions, SIAM J. Control Optim. 29 (1991), 474–492. [CaF] Cannarsa, P. and Frankowska, H., Value function and optimality condition for semilinear control problems. II. The parabolic case, Appl. Math. Optim. 33 (1996), 1-33. [C1] Clarke, F. H., Necessary Conditions for Nonsmooth Problems in Optimal Control and the Calculus of Variations, Doctoral thesis, University of Washington, 1973. 32] Clarke, F. H., Generalized gradients and applications, Trans. Amer. Math. Soc. 205 (1975), 247-262. [C3] Clarke, F. H., The maximum principle under minimal hypotheses, SIAM J. Control Optim. 14 (1976), 1078-1091. [C4] Clarke, F. H., Optimization and Nonsmooth Analysis, Wiley Interscience, New York, 1983; reprinted as vol. 5 of Classics in Applied Mathematics, SIAM, Philadelphia, PA, 1990; Russian translation, Nauka, Moscow, 1988. [C5] Clarke, F. H., Methods of Dynamic and Nonsmooth Optimization, CBMS/NSF Regional Conf. Ser. in Appl. Math., vol. 57, SIAM, Philadelphia, PA, 1989. [C6] Clarke, F. H., An indirect method in the calculus of variations, Trans. Amer. Math. Soc. 336 (1993), 655-673. [CL1] Clarke, F. H. and Ledyaev, Yu., Mean value inequalities, Proc. Amer. Math. Soc. 122 (1994), 1075-1083. [CL2] Clarke, F. H. and Ledyaev, Yu., Mean value inequalities in Hilbert space, Trans. Amer. Math. Soc. 344 (1994), 307-324. [CL] Clarke, F. H. and Loewen, P. D., The value function in optimal control: Sensitivity, controllability and time-optimality, SIAM J. Control Optim. 24 (1986), 243-263. [CLS1] ] Clarke, F. H., Ledyaev, Yu., and Stern, R. J., Fixed points and equilibria in nonconvex sets, Nonlinear Anal. 25 (1995), 145–161. [CLS2] Clarke, F. H., Ledyaev, Yu., and Stern, R. J., Asymptotic stability and smooth Lyapounov functions, preprint, 1997. [CLSS] Clarke, F. H., Ledyaev, Yu., Sontag, E. D., and Subbotin, A. I., Asymptotic controllability implies feedback stabilization, IEEE Trans. Automat. Control, in press (1997). [CLSW] 1] Clarke, F. H., Ledyaev, Yu., Stern, R. J., and Wolenski, P. R., Qualitative properties of trajectories of control systems: A survey, J. Dynam. Control Systems 1 (1995), 1-47. [CLW] Clarke, F. H., Ledyaev, Yu., and Wolenski, P. R., Proximal analysis and minimization principles, J. Math. Anal. Appl. 196 (1995), 722-735. \( \left\lbrack \mathrm{{CR}}\right\rbrack \) Clarke, F. H. and Redheffer, R., The proximal subgradient and constancy, Canad. Math. Bull. 36 (1993), 30-32. [CSW] Clarke, F. H., Stern, R. J., and Wolenski, P. R., Subgradient criteria for Lipschitz behavior, monotonicity and convexity, Canad. J. Math. 45 (1993), 1167-1183. CV] Clarke, F. H. and Vinter, R. B., Local optimality conditions and Lipschitzian solutions to the Hamilton-Jacobi equations, SIAM J. Control Optim. 21 (1983), 856-870. [CPT] Coron, J.-M., Praly, L., and Teel, A., Feedback stabilization of nonlinear systems: Sufficient conditions and Lyapounov and input-output techniques, In Trends in Control (A. Isidori, Ed.), Springer-Verlag, New York, 1995. Cx] Cox, S. J., The shape of the ideal column, Math. Intelligencer 14 (1) (1992), 16-24. [CO] Cox, S. J. and Overton, M. L., On the optimal design of columns against buckling, SIAM J. Math. Anal. 23 (1992), 287-325. [CIL] Crandall, M. G., Ishii, H., and Lions, P.-L., User's guide to viscosity solutions of second-order partial differential equations, Bull. Amer. Math. Soc. 27 (1992), 1–67. [D] Deimling, K., Multivalued Differential Equations, de Gruyter, Berlin, 1992. [DR] Demyanov, V. F. and Rubinov, A. M., Constructive Nonsmooth Analysis, Peter Lang, Frankfurt, 1995. [DGZ] Deville, R., Godefroy, G., and Zizler, V., Smoothness and Renorm-ings in Banach Spaces, Longman and Wiley, New York, 1993. [E] Ekeland, I., Nonconvex minimization problems, Bull. Amer. Math. Soc. (New Series) 1 (1979), 443-474. [F] Filippov, A. F., Differential Equations with Discontinuous RightHand Sides, Kluwer Academic, Dordrecht, 1988. [FS] Fleming, W. H. and Soner, H. M., Controlled Markov Processes and Viscosity Solutions, Springer-Verlag, New York, 1993. [Fr] Frankowska, H., Lower semicontinuous solutions of the Hamilton-Jacobi equation, SIAM J. Control Optim. 31 (1993), 257-272. [FK] Freeman, R. A. and Kokotovic, P. V., Robust Nonlinear Control Design. State-Space and Lyapounov Techniques, Birkhäuser, Boston, 1996. [H] Havelock, D., A Generalization of the Hamilton-Jacobi Equation, Master's thesis, University of British Columbia, Vancouver, Canada, 1977. [He] Hermes, H., Resonance, stabilizing feedback controls, and regularity of viscosity solutions of Hamilton-Jacobi-Bellman equations, Math. Control Signals Systems 9 (1996), 59–72. [HL] Hermes, H. and Lasalle, J. P., Functional Analysis and Time Optimal Control, Academic Press, New York, 1969. [HS] Heymann, M. and Stern, R. J., \( \Omega \) rest points in autonomous control systems, J. Differential Equations 20 (1976), 389-398. [UL] Hiriart-Urruty, J.-B. and Lemaréchal, C., Convex Analysis and Minimization Algorithms. I. Fundamentals, Springer-Verlag, Berlin, 1993. [HY] Hiriart-Urruty, J.-B. and Ye, D., Sensitivity analysis of all eigenvalues of a symmetric matrix, Numer. Math. 70 (1995), 45-72. Ioffe, A. D., Proximal analysis and approximate subdifferentials, J. London M
1012_(GTM178)Nonsmooth Analysis and Control Theory
94
gn. State-Space and Lyapounov Techniques, Birkhäuser, Boston, 1996. [H] Havelock, D., A Generalization of the Hamilton-Jacobi Equation, Master's thesis, University of British Columbia, Vancouver, Canada, 1977. [He] Hermes, H., Resonance, stabilizing feedback controls, and regularity of viscosity solutions of Hamilton-Jacobi-Bellman equations, Math. Control Signals Systems 9 (1996), 59–72. [HL] Hermes, H. and Lasalle, J. P., Functional Analysis and Time Optimal Control, Academic Press, New York, 1969. [HS] Heymann, M. and Stern, R. J., \( \Omega \) rest points in autonomous control systems, J. Differential Equations 20 (1976), 389-398. [UL] Hiriart-Urruty, J.-B. and Lemaréchal, C., Convex Analysis and Minimization Algorithms. I. Fundamentals, Springer-Verlag, Berlin, 1993. [HY] Hiriart-Urruty, J.-B. and Ye, D., Sensitivity analysis of all eigenvalues of a symmetric matrix, Numer. Math. 70 (1995), 45-72. Ioffe, A. D., Proximal analysis and approximate subdifferentials, J. London Math. Soc. 41 (1990), 175-192. [Is] Isidori, A., Nonlinear Control Systems: An Introduction, Springer-Verlag, Berlin, 1985. [JDT] Jofré, A., Dihn, T. L., and Théra, M., \( \varepsilon \) -Subdifferential calculus for nonconvex functions and \( \varepsilon \) -monotonicity, C. R. Acad. Sci. Paris Sér. I Math. 323 (1996), 735–740. [KS] Krasovskii, N. N. and Subbotin, A. I., Game-Theoretical Control Problems, Springer-Verlag, New York, 1988. 270 Bibliography [Kry] Kryazhimskii, A., Optimization of the ensured result for dynamical systems, Proceedings of the International Congress of Mathematicians, vols. 1, 2 (Berkeley, CA, 1986), pp. 1171-1179, American Mathematical Society, Providence, RI, 1987. LM] Lakshmikantham, V. Leela, S., and Martynyuk, A., Practical Stability of Nonlinear Systems, World Scientific, Singapore, 1990. [Le] Leitmann, G., One approach to the control of uncertain dynamical systems, Appl. Math. Comput. 70 (1995), 261-272. [LS] Li, Y. and Shi, S., A generalization of Ekeland’s \( \varepsilon \) -variational principle and of its Borwein-Preiss smooth variant, J. Math. Anal. Appl., to appear. [L] Lions, J. L., Contrôle optimal de systèmes gouvernés par des équations aux dérivées partielles, Dunod, Paris, 1968. [L1] Loewen, P. D., Optimal Control via Nonsmooth Analysis, CRM Proc. Lecture Notes, vol. 2, American Mathematical Society, Providence, RI, 1993. [L2] Loewen, P. D., A mean value theorem for Fréchet subgradients, Nonlinear Anal. 23 (1994), 1365–1381. [Ly] Lyusternik, L. A., The Shortest Lines-Variational Problems, MIR, Moscow, 1976. [MN] Mäkelä, M. M. and Neittaanmäki, P., Nonsmooth Optimization, World Scientific, London, 1992. [MW] Mawhin, J. and Willem, M., Critical Point Theory and Hamiltonian Systems, Springer-Verlag, New York, 1989. [MS] Mordukhovich, B. S. and Shao, Y. H., On nonconvex subdifferen-tial calculus in Banach spaces, J. Convex Anal. 2 (1995), 211-227. [O] Overton, M. L. Large-scale optimization of eigenvalues, SIAM J. Optim. 2 (1992), 88–120. [Pa] Panagiotopoulos, P. D., Hemivariational Inequalities. Applications in Mechanics and Engineering, Springer-Verlag, Berlin, 1993. [PQ] Penot, J.-P. and Quang, P. H., Generalized convexity of functions and generalized monotonicity of set-valued maps, J. Optim. Theory Appl. \( \mathbf{{92}}\left( {1997}\right) ,{343} - {356} \) . [Ph] Phelps, R. R., Convex Functions, Monotone Operators and Differentiability, 2nd ed., Lecture Notes in Mathematics, vol. 1364, Springer-Verlag, New York, 1993. o] Polak, E.,, On the mathematical foundations of nondifferentiable optimization in engineering design, SIAM Rev. 29 (1987), 21-89. [PBGM] Pontryagin, L. S., Boltyanskii, R. V., Gamkrelidze, R. V., and Mischenko, E. F., The Mathematical Theory of Optimal Processes, Wiley, New York, 1962. [RW] Redheffer, R. and Walter, W., The subgradient in \( {\mathbb{R}}^{n} \), Nonlinear Anal. 20 (1993), 1345–1348. [R1] Rockafellar, R. T., Convex Analysis, Princeton Mathematical Series, vol. 28, Princeton University Press, Princeton, NJ, 1970. [R2] Rockafellar, R. T., The Theory of Subgradients and Its Applications to Problems of Optimization: Convex and Nonconvex Functions, Helderman Verlag, Berlin, 1981. [R3] Rockafellar, R. T., Lagrange multipliers and optimality, SIAM Rev. 35 (1993), 183-238. [RolV] Roland, J. D. and Vinter, R. B., Construction of optimal feedback control, System Control Lett. 16 (1991), 357-367. [Ro] Roxin, E. O., Control Theory and its Applications, Gordon and Breach, New York, 1996. Rudin, W., Real and Complex Analysis, McGraw-Hill, New York, 1966. Rudin, W., Functional Analysis, McGraw-Hill, New York, 1973. \( \left\lbrack \mathrm{{Sa}}\right\rbrack \) Saks, S., Theory of the Integral, Monografie Matematyczne Ser., no. 7 (1937); 2nd rev. ed., Dover, New York, 1964. Sontag, E. D., Mathematical Control Theory, Texts in Applied Mathematics vol. 6, Springer-Verlag, New York, 1990. [Su] Subbotin, A. I., Generalized Solutions of First-Order PDEs, Birkhäuser, Boston, 1995. [SSY] Sussmann, H., Sontag, E., and Yang, Y., A general result on the stabilization of linear systems using bounded controls, IEEE Trans. Automat. Control 39 (1994), 2411-2425. [T1] Tikhomirov, V. M., Stories about Maxima and Minima, American Mathematical Society, Providence, RI, 1990. 272 Bibliography [T2] Tikhomirov, V. M., Convex Analysis and Approximation Theory, Analysis II, Encyclopaedia of Mathematical Sciences (R. V. Gamkrelidze, Ed.), vol. 14, Springer-Verlag, New York, 1990. Veliov, V., Sufficient conditions for viability under imperfect measurement, Set-Valued Anal. 1 (1993), 305-317. [Vi] Vinter, R. B., forthcoming monograph. [W] Weckesser, V., The subdifferential in Banach spaces, Nonlinear Anal. 20 (1993), 1349–1354. [WZ] Wolenski, P. R. and Zhuang, Y., Proximal analysis and the minimal time function, SIAM J. Control Optim., to appear. [Y] Young, L. C., Lectures on the Calculus of Variations and Optimal Control Theory, Saunders, Philadelphia, PA, 1969. Zabczyk, J., Mathematical Control Theory: An Introduction, Birkhäuser, Boston, 1992. ## Index abnormal, 107, 240 absolutely continuous, 162, 177 almost everywhere solution, 227 approximation by \( P \) -subgradients, 138 attainability, 215 attainable set, 193 augmented Hamiltonian, 223 Aumann, 157 autonomous, 190 Borwein and Preiss, 43 Bouligand tangent cone, 8, 90 Brouwer's theorem, 202, 249 calculus of variations, 162, 163 calm, 170 Carathéodory, 223 chain rule, 32, 48, 58, 76 closed-valued, 150 closest point, 22 compactness of approximate tra- jectories, 185 comparison theorems, 225 constancy of the Hamiltonian, 241 constrained optimization, 49, 103 constraint qualification, 128 contingent, 90 continuous function, 28 continuously differentiable, 33 controllability, 244 convex functions, 29, 51, 80 convex set, 26 \( D \) -normal cone,141 \( D \) -subdifferential,138 \( D \) -subgradient,138 \( D \) -tangent cone,141 Danskin's theorem, 99 decoupling, 55 decrease principle, 122, 171 decreasing function, 2 density theorem, 39 dependence on initial conditions, 201 derivate, 2, 136 diameter, 181 differential inclusion, 177 Dini, vii, 2 Dini subderivate, 98 direction of descent, 97 directional calculus, 139 directional derivative, 31 directional subderivate, 4 directional subgradient, 138 distance function, 8, 23 domain, 27 Dubois-Reymond lemma, 162 dynamic feedback, 215 eigenvalue, 10 endpoint cost, 222 epigraph, 28 equilibria, 202 equilibrium, 1, 202 Erdmann condition, 164 Erdmann transform, 241 Euler equation, 163 Euler inclusion, 163 Euler polygonal arc, 181 Euler solution, 181 Euler solutions, 180 exact penalization, 50 extended real-valued, 27 feasible set, 104 feedback, 192, 196, 229 feedback selections, 192 feedback synthesis, 228 Fermat's rule, 2 Filippov's lemma, 174, 178 fixed point theorem, 202 flow-invariant, 7 Fréchet derivative, 31 Fréchet differentiable, 31, 148 free time problems, 243 Fritz John necessary conditions, 100 fuzzy sum rule, 56 Gâteaux differentiable, 31 generalized directional derivative, 70 generalized gradient, 6, 69, 72, 160 generalized gradient formula, 93 generalized Jacobian, 108, 133 <table><tr><td>global null controllability, 246</td></tr><tr><td>globally asymptotically stable, 208 gradient, 104</td></tr><tr><td>graph, 28, 150</td></tr><tr><td>Grave–Lyusternik, 127</td></tr><tr><td>Gronwall’s lemma, 179</td></tr><tr><td>growth, 209, 210</td></tr><tr><td>growth hypothesis, 105</td></tr><tr><td>Gâteaux derivative, 78</td></tr><tr><td>half-space, 27 Hamilton–Jacobi equation, 17, 224 Hamilton–Jacobi inequality, 222, 228</td></tr><tr><td>Hamiltonian function, 16</td></tr><tr><td>Hamiltonian inclusion, 231, 255</td></tr><tr><td>Hausdorff continuous, 175</td></tr><tr><td>Hausdorff distance, 175 Hessian, 32</td></tr><tr><td>homeomorphic, 203</td></tr><tr><td>horizontal approximation theorem. 67</td></tr><tr><td>hyperplane, 27</td></tr><tr><td>implicit and inverse functions, 108, 129</td></tr><tr><td>indicator function, 28</td></tr><tr><td>inf-convolution, 44</td></tr><tr><td>infinitesimal decrease, 209, 210</td></tr><tr><td>initial-value problem, 180</td></tr><tr><td>inner product, 18</td></tr><tr><td>integral functionals, 148 invariance, 188</td></tr><tr><td>invariant embedding, 223</td></tr><tr><td>inverse function theorem, 2</td></tr><tr><td>inverse functions, 133</td></tr><tr><td>Jacobian, 2, 104, 106</td></tr><tr><td>Kakutani, 249</td></tr><tr><td>\( {L}^{1} \) -optimization, \( {11} \)</td></tr><tr><td>Lagrange Multiplier Rule, 15, 65, 104</td></tr><tr><td>Lagrangian, 110</td></tr></table> Lebourg's Mean Value Theorem, 75 Legendre, 222 limiting calculus, 61 limiting chain rule, 65 limiting normal, 62 limiting subdifferential, 61, 160 Line-Fitting, 98 linear growth, 178 linearization, 1 Lipschitz, 39, 51, 52 Lipschitz Inverse Function Theo- rem, 135
1012_(GTM178)Nonsmooth Analysis and Control Theory
95
em. 67</td></tr><tr><td>hyperplane, 27</td></tr><tr><td>implicit and inverse functions, 108, 129</td></tr><tr><td>indicator function, 28</td></tr><tr><td>inf-convolution, 44</td></tr><tr><td>infinitesimal decrease, 209, 210</td></tr><tr><td>initial-value problem, 180</td></tr><tr><td>inner product, 18</td></tr><tr><td>integral functionals, 148 invariance, 188</td></tr><tr><td>invariant embedding, 223</td></tr><tr><td>inverse function theorem, 2</td></tr><tr><td>inverse functions, 133</td></tr><tr><td>Jacobian, 2, 104, 106</td></tr><tr><td>Kakutani, 249</td></tr><tr><td>\( {L}^{1} \) -optimization, \( {11} \)</td></tr><tr><td>Lagrange Multiplier Rule, 15, 65, 104</td></tr><tr><td>Lagrangian, 110</td></tr></table> Lebourg's Mean Value Theorem, 75 Legendre, 222 limiting calculus, 61 limiting chain rule, 65 limiting normal, 62 limiting subdifferential, 61, 160 Line-Fitting, 98 linear growth, 178 linearization, 1 Lipschitz, 39, 51, 52 Lipschitz Inverse Function Theo- rem, 135 Lipschitz near \( x,{39} \) local attainable, 220 locally attainable, 220, 249 locally Lipschitz, 39, 196 locally Lipschitz selection, 249 lower Hamiltonian, 188 lower semicontinuity, 28, 167 Lyapounov functions, 208 Lyapounov pair, 209 manifold, 26 maximal rank, 244 maximum principle, 237 mean value inequality, 111, 117 Mean Value Theorem, 31, 111 measurable multifunctions, 149 measurable selections, 149, 151 mesh size, 181 method of characteristics, 255 metric projection, 192 Metzler matrix, 249 minimal time function, 16, 246 minimal time problem, 254 minimax problems, 10 minimax solutions, 225 minimax theorem, 234 minimization principles, 21, 43 monotonicity, 65, 209 Moreau–Yosida approximation, 64 multidirectionality, 112 multifunction, 177 multiplier, 15, 110 multiplier rule, 15, 104 multiplier set, 106, 107 necessary conditions, 230 nonautonomous case, 200 normal, 107, 244 normal cone, 85 normality, 244 null controllability, 244 open ball, 18 optimal control problem, 222 orthogonal decomposition, 25 partial proximal subdifferential, 38 partial proximal subgradients, 38 partial subdifferentials, 143 partial subgradients, 66 partition of unity, 205 pointed, 168 polar, 65, 85, 168 Pontryagin's Maximum Principle, 237 Pontryagin's maximum principle, 252 positive definiteness, 209, 210 positively homogeneous, 70 principle of optimality, 16, 223 projection, 9, 22 proximal aiming, 189 proximal Gronwall inequality, 68 proximal Hamilton-Jacobi equa- tion, 224 proximal mean value theorem, 64 proximal normal cone, 22 proximal normal direction, 9 proximal normal inequality, 25 proximal normals, 21 proximal solution, 224 proximal subdifferential, 5, 29 proximal subgradient, 21, 29 proximal subgradient inequality, 33 proximal superdifferential, 37 quadratic inf-convolutions, 44 276 Index Rademacher’s Theorem, 6, 93 Rayleigh's formula, 11 reachable set, 251 regular, 69, 81, 91 rest point, 202 selection, 180 semigroup property, 193 semisolution, 228 sensitivity, 105 separation theorem, 27 smooth manifold, 7 solution, 177 solvability, 105, 126 stabilizing feedback, 214 standard control system, 178 Standing Hypotheses, 178 state augmentation, 200 static feedback, 215 Stegall's minimization principle, 43 strictly convex, 252 strictly differentiable, 96 strong invariance, 195, 198 strong weak monotonicity, 123 strongly decreasing, 123, 124, 217, 219, 251 strongly increasing, 219, 251 strongly invariant, 198 subadditive, 70 Subbotin, 137 subderivate, 136 sum rule, 38, 54 support function, 71, 124 tangency, 69 tangent, 83 tangent cone, 84 tangent space, 7 terminal constraints, 238 the origin is normal, 244 Torricelli point, 12 Torricelli's Table, 11 trajectory, 178 trajectory continuation, 187 transversality condition, 164, 231 unilateral constraint set, 149 upper envelopes, 65 upper Hamiltonian, 188 upper semicontinuous, 28, 73, 179 value function, 104, 224 value function analysis, 238 variable intervals, 187 variational problem, 162 verification functions, 222 viability, 191 viscosity solutions, 142, 226 weak invariance, 188, 190, 192 weak sequential compactness, 164 weakly avoidable, 248 weakly decreasing, 124, 211, 215, 251 weakly increasing, 251 weakly lower semicontinuous, 55 weakly predecreasing, 251 weakly preincreasing, 251 weakly preinvariant, 194 wedged, 166 zero of a multifunction, 202 ## Graduate Texts in Mathematics 61 Whitehead. Elements of Homotopy Theory. 62 Kargapolov/Merlzjakov. Fundamentals of the Theory of Groups. 63 Bollobas. Graph Theory. 64 Edwards. Fourier Series. Vol. I 2nd ed. 65 WELLs. Differential Analysis on Complex Manifolds. 2nd ed. 66 Waterhouse. Introduction to Affine Group Schemes. 67 Serre. Local Fields. 68 WEIDMANN. Linear Operators in Hilbert Spaces. 69 Lang. Cyclotomic Fields II. 70 Massey. Singular Homology Theory. 71 FARKAS/KRA. Riemann Surfaces. 2nd ed. 72 Stillwell. Classical Topology and Combinatorial Group Theory. 2nd ed. 73 Hungerford. Algebra. 74 Davenport. Multiplicative Number Theory. 2nd ed. 75 HOCHSCHILD. Basic Theory of Algebraic Groups and Lie Algebras. 76 Ittaka. Algebraic Geometry. 77 HECKE. Lectures on the Theory of Algebraic Numbers. 78 Burris/Sankappanavar. A Course in Universal Algebra. 79 WALTERS. An Introduction to Ergodic Theory. 80 Robinson. A Course in the Theory of Groups. 2nd ed. 81 Forster. Lectures on Riemann Surfaces. 82 Bott/Tu. Differential Forms in Algebraic Topology. 83 WASHINGTON. Introduction to Cyclotomic Fields. 2nd ed. 84 IRELAND/ROSEN. A Classical Introduction to Modern Number Theory. 2nd ed. 85 Edwards. Fourier Series. Vol. II. 2nd ed. 86 VAN LINT. Introduction to Coding Theory. 2nd ed. 87 Brown. Cohomology of Groups. 88 Pierce. Associative Algebras. 89 LANG. Introduction to Algebraic and Abelian Functions. 2nd ed. 90 BRØNDSTED. An Introduction to Convex Polytopes. 91 BEARDON. On the Geometry of Discrete Groups. 92 Diestel. Sequences and Series in Banach Spaces. 93 Dubrovin/Fomenko/Novikov. Modem Geometry-Methods and Applications. Part I. 2nd ed. 94 WARNER. Foundations of Differentiable Manifolds and Lie Groups. 95 Shiryaev. Probability. 2nd ed. 96 Conway. A Course in Functional Analysis. 2nd ed. 97 KOBLITZ. Introduction to Elliptic Curves and Modular Forms. 2nd ed. 98 BRÖCKER/TOM DIECK. Representations of Compact Lie Groups. 99 Grove/Benson. Finite Reflection Groups. 2nd ed. 100 Berg/Christensen/Ressel. Harmonic Analysis on Semigroups: Theory of Positive Definite and Related Functions. 101 Edwards. Galois Theory. 102 VaradaranaN. Lie Groups, Lie Algebras and Their Representations. 103 LANG. Complex Analysis. 3rd ed. 104 Dubrovin/FOMENKO/Novikov. Modern Geometry-Methods and Applications. Part II. 105 LANG. \( S{L}_{2}\left( \mathbf{R}\right) \) . 106 SILVERMAN. The Arithmetic of Elliptic Curves. 107 OLVER. Applications of Lie Groups to Differential Equations. 2nd ed. 108 RANGE. Holomorphic Functions and Integral Representations in Several Complex Variables. 109 LEHTO. Univalent Functions and Teichmüller Spaces. 110 LANG. Algebraic Number Theory. 111 Husemöller. Elliptic Curves. 112 LANG. Elliptic Functions. 113 KARATZAS/SHREVE. Brownian Motion and Stochastic Calculus. 2nd ed. 114 KOBLITZ. A Course in Number Theory and Cryptography. 2nd ed. 115 BERGER/GOSTIAUX. Differential Geometry: Manifolds, Curves, and Surfaces. 116 Kelley/Srinivasan. Measure and Integral. Vol. I. 117 Serre. Algebraic Groups and Class Fields. 118 Pedersen. Analysis Now. 119 ROTMAN. An Introduction to Algebraic Topology. 120 ZIEMER. Weakly Differentiable Functions: Sobolev Spaces and Functions of Bounded Variation. 121 LANG. Cyclotomic Fields I and II. Combined 2nd ed. 122 REMMERT. Theory of Complex Functions. Readings in Mathematics 123 EBBINGHAUS/HERMES et al. Numbers. Readings in Mathematics 124 DUBROVIN/FOMENKO/NOVIKOV. Modem Geometry-Methods and Applications. Part III. 125 Berenstein/Gay. Complex Variables: An Introduction. 126 BOREL. Linear Algebraic Groups. 2nd ed. 127 MASSEY. A Basic Course in Algebraic Topology. 128 Rauch. Partial Differential Equations. 129 FULTON/HARRIS. Representation Theory: A First Course. Readings in Mathematics 130 Dodson/Poston. Tensor Geometry. 131 LAM. A First Course in Noncommutative Rings. 132 BEARDON. Iteration of Rational Functions. 133 Harris. Algebraic Geometry: A First Course. 134 ROMAN. Coding and Information Theory. 135 ROMAN. Advanced Linear Algebra. 136 Adkins/Weintraub. Algebra: An Approach via Module Theory. 137 AXLER/BOURDON/RAMEY. Harmonic Function Theory. 138 COHEN. A Course in Computational Algebraic Number Theory. 139 BREDON. Topology and Geometry. 140 AUBIN. Optima and Equilibria. An Introduction to Nonlinear Analysis. 141 BECKER/WEISPFENNING/KREDEL. Gröbner Bases. A Computational Approach to Commutative Algebra. 142 LANG. Real and Functional Analysis. 3rd ed. 143 DOOB. Measure Theory. 144 DENNIS/FARB. Noncommutative Algebra. 145 VICK. Homology Theory. An Introduction to Algebraic Topology. 2nd ed. 146 BRIDGES. Computability: A Mathematical Sketchbook. 147 ROSENBERG. Algebraic \( K \) -Theory and Its Applications. 148 ROTMAN. An Introduction to the Theory of Groups. 4th ed. 149 RATCLIFFE. Foundations of Hyperbolic Manifolds. 150 EISENBUD. Commutative Algebra with a View Toward Algebraic Geometry. 151 SILVERMAN. Advanced Topics in the Arithmetic of Elliptic Curves. 152 ZIEGLER. Lectures on Polytopes. 153 FULTON. Algebraic Topology: A First Course. 154 BROWN/PEARCY. An Introduction to Analysis. 155 KASSEL. Quantum Groups. 156 KECHRIS. Classical Descriptive Set Theory. 157 MALLIAVIN. Integration and Probability. 158 ROMAN. Field Theory. 159 CONWAY. Functions of One Complex Variable II. 160 LANG. Differential and Riemannian Manifolds. 161 BORWEIN/ERDÉLYI. Polynomials and Polynomial Inequalities. 162 ALPERIN
1012_(GTM178)Nonsmooth Analysis and Control Theory
96
Analysis. 3rd ed. 143 DOOB. Measure Theory. 144 DENNIS/FARB. Noncommutative Algebra. 145 VICK. Homology Theory. An Introduction to Algebraic Topology. 2nd ed. 146 BRIDGES. Computability: A Mathematical Sketchbook. 147 ROSENBERG. Algebraic \( K \) -Theory and Its Applications. 148 ROTMAN. An Introduction to the Theory of Groups. 4th ed. 149 RATCLIFFE. Foundations of Hyperbolic Manifolds. 150 EISENBUD. Commutative Algebra with a View Toward Algebraic Geometry. 151 SILVERMAN. Advanced Topics in the Arithmetic of Elliptic Curves. 152 ZIEGLER. Lectures on Polytopes. 153 FULTON. Algebraic Topology: A First Course. 154 BROWN/PEARCY. An Introduction to Analysis. 155 KASSEL. Quantum Groups. 156 KECHRIS. Classical Descriptive Set Theory. 157 MALLIAVIN. Integration and Probability. 158 ROMAN. Field Theory. 159 CONWAY. Functions of One Complex Variable II. 160 LANG. Differential and Riemannian Manifolds. 161 BORWEIN/ERDÉLYI. Polynomials and Polynomial Inequalities. 162 ALPERIN/BELL. Groups and Representations. 163 DIXON/MORTIMER. Permutation Groups. 164 Nathanson. Additive Number Theory: The Classical Bases. 165 NATHANSON. Additive Number Theory: Inverse Problems and the Geometry of Sumsets. 166 SHARPE. Differential Geometry: Cartan's Generalization of Klein's Erlangen Program. 167 MORANDI. Field and Galois Theory. 168 Ewald. Combinatorial Convexity and Algebraic Geometry. 169 BHATIA. Matrix Analysis. 170 BREDON. Sheaf Theory. 2nd ed. 171 Petersen. Riemannian Geometry. 172 REMMERT. Classical Topics in Complex Function Theory. 173 Diestel. Graph Theory. 174 BRIDGES. Foundations of Real and Abstract Analysis. 175 LICKORISH. An Introduction to Knot Theory. 176 Lee. Riemannian Manifolds. 177 Newman. Analytic Number Theory. 178 Clarke/Ledyaev/Stern/Wolenski. Nonsmooth Analysis and Control Theory.
1013_(GTM179)Banach Algebra Techniques in Operator Theory
0
# GraduateTexts inMathematics Ronald G. Douglas # Banach Algebra Techniques in Operator Theory Second Edition Springer ## Graduate Texts in Mathematics 179 Editorial Board S. Axler F.W. Gehring K.A. Ribet Springer Science+Business Media, LLC ## Graduate Texts in Mathematics 1 TAKEUTI/ZARING. Introduction to Axiomatic Set Theory. 2nd ed. 2 Oxtoby. Measure and Category. 2nd ed. 3 Schaefer. Topological Vector Spaces. 4 Hilton/Stammbach. A Course in Homological Algebra. 2nd ed. 5 MAC LANE. Categories for the Working Mathematician. 2nd ed. 6 Hughes/Piper. Projective Planes. 7 Serre. A Course in Arithmetic. 8 TAKEUTI/ZARING. Axiomatic Set Theory. 9 Humphreys. Introduction to Lie Algebras and Representation Theory. 10 Cohen. A Course in Simple Homotopy Theory. 11 Conway. Functions of One Complex Variable I. 2nd ed. 12 Beals. Advanced Mathematical Analysis. 13 Anderson/Fuller. Rings and Categories of Modules. 2nd ed. 14 Golubitsky/Guillemin. Stable Mappings and Their Singularities. 15 Berberlan. Lectures in Functional Analysis and Operator Theory. 16 WINTER. The Structure of Fields. 17 Rosenblatt. Random Processes. 2nd ed. 18 Halmos. Measure Theory. 19 Halmos. A Hilbert Space Problem Book. 2nd ed. 20 Husemoller. Fibre Bundles. 3rd ed. 21 Humphreys. Linear Algebraic Groups. 22 Barnes/Mack. An Algebraic Introduction to Mathematical Logic. 23 Greub. Linear Algebra. 4th ed. 24 Holmes. Geometric Functional Analysis and Its Applications. 25 Hewitt/Stromberg. Real and Abstract Analysis. 26 Manes. Algebraic Theories. 27 Kelley. General Topology. 28 ZARISKI/SAMUEL. Commutative Algebra. Vol.I. 29 Zariski/Samuel. Commutative Algebra. Vol.II. 30 JACOBSON. Lectures in Abstract Algebra I. Basic Concepts. 31 JACOBSON. Lectures in Abstract Algebra II. Linear Algebra. 32 JACOBSON. Lectures in Abstract Algebra III. Theory of Fields and Galois Theory. 33 Hirsch. Differential Topology. 34 Sprtzer. Principles of Random Walk. 2nd ed. 35 ALEXANDER/WERMER. Several Complex Variables and Banach Algebras. 3rd ed. 36 Kelley/Namioka et al. Linear Topological Spaces. 37 Monk. Mathematical Logic. 38 Grauert/Fritzsche. Several Complex Variables. 39 Arveson. An Invitation to \( {C}^{ * } \) -Algebras. 40 Kemeny/Snell/Knapp. Denumerable Markov Chains. 2nd ed. 41 Apostol. Modular Functions and Dirichlet Series in Number Theory. 2nd ed. 42 Serre. Linear Representations of Finite Groups. 43 GILLMAN/JERISON. Rings of Continuous Functions. 44 Kendig. Elementary Algebraic Geometry. 45 Loëve. Probability Theory I. 4th ed. 46 Loëve. Probability Theory II. 4th ed. 47 MoISE. Geometric Topology in Dimensions 2 and 3. 48 SACHS/WU. General Relativity for Mathematicians. 49 Gruenberg/Weir. Linear Geometry. 2nd ed. 50 EDWARDS. Fermat's Last Theorem. 51 KLINGENBERG. A Course in Differential Geometry. 52 Hartshorne. Algebraic Geometry. 53 Mann. A Course in Mathematical Logic. 54 Graver/Watkins. Combinatorics with Emphasis on the Theory of Graphs. 55 Brown/Pearcy. Introduction to Operator Theory I: Elements of Functional Analysis. 56 Massey. Algebraic Topology: An Introduction. 57 Crowell/Fox. Introduction to Knot Theory. 58 KOBLITZ. \( p \) -adic Numbers, \( p \) -adic Analysis, and Zeta-Functions. 2nd ed. 59 LANG. Cyclotomic Fields. 60 Arnold. Mathematical Methods in Classical Mechanics. 2nd ed. Ronald G. Douglas # Banach Algebra Techniques in Operator Theory Second Edition Ronald G. Douglas Department of Mathematics Texas A & M University College Station, TX 77843-1248 USA Editorial Board S. Axler F.W. Gehring K.A. Ribet Mathematics Department Mathematics Department Department of Mathematics San Francisco State University of Michigan East Hall University of University Ann Arbor, MI 48109 California at Berkeley San Francisco, CA 94132 USA Berkeley, CA 94720-3840 USA USA ## Mathematics Subject Classification (1991): 46Bxx, 46Jxx, 46Lxx, 47Bxx Library of Congress Cataloging-in-Publication Data Douglas, Ronald G. Banach algebra techniques in operator theory / Ronald G. Douglas. - 2nd ed. p. cm. - (Graduate texts in mathematics ; 179) Includes bibliographical references (p. - ) and index. ISBN 978-0-387-98377-6 ISBN 978-1-4612-1656-8 (eBook) DOI 10.1007/978-1-4612-1656-8 1. Operator theory. 2. Banach algebras. 3. Hilbert space I. Title. II. Series. QA329.D68 1998 97-42889 515'.724-dc21 Printed on acid-free paper. The first edition of this book was published by Academic Press, New York, (c) 1972. (C) 1998 Springer Science+Business Media New York Originally published by Springer-Verlag New York, Inc. in 1998 All rights reserved. This work may not be translated or copied in whole or in part without the written permission of the publisher Springer Science+Business Media, LLC, except for brief excerpts in connection with reviews or scholarly analysis. Use in connection with any form of information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed is forbidden. The use of general descriptive names, trade names, trademarks, etc., in this publication, even if the former are not especially identified, is not to be taken as a sign that such names, as understood by the Trade Marks and Merchandise Marks Act, may accordingly be used freely by anyone. Production managed by Timothy Taylor; manufacturing supervised by Jeffrey Taub. Typeset by BookComp, Inc., Grand Rapids, MI. ## Preface to the Second Edition In the quarter century since the first edition of this book appeared, tremendous development has occurred in operator theory and the topics covered here. However, the new edition remains unchanged except that several mistakes and typographical errors have been corrected. Further, a brief report on the current state of the double-asterisk, open, problems is given along with references. No attempt is made to describe other progress that has been made in the study of Toeplitz operators and related topics nor has the bibliography been updated. Still, it is hoped that a new generation of students will find useful the introduction to operator theory given here. College Station, Texas Ronald G. Douglas July 1997 ## Preface to the First Edition Operator theory is a diverse area of mathematics which derives its impetus and motivation from several sources. It began as did practically all of modern analysis with the study of integral equations at the end of the last century. It now includes the study of operators and collections of operators arising in various branches of physics and mechanics as well as other parts of mathematics and indeed is sufficiently well developed to have a logic of its own. The appearance of several monographs on recent studies in operator theory testifies both to its vigor and breadth. The intention of this book is to discuss certain advanced topics in operator theory and to provide the necessary background for them assuming only the standard senior-first year graduate courses in general topology, measure theory, and algebra. There is no attempt at completeness and many "elementary" topics are either omitted or mentioned only in the problems. The intention is rather to obtain the main results as quickly as possible. The book begins with a chapter presenting the basic results in the theory of Banach spaces along with many relevant examples. The second chapter concerns the elementary theory of commutative Banach algebras since these techniques are essential for the approach to operator theory presented in the later chapters. Then after a short chapter on the geometry of Hilbert space, the study of operator theory begins in earnest. In the fourth chapter operators on Hilbert space are studied and a rather sophisticated version of the spectral theorem is obtained. The notion of a \( {C}^{ * } \) -algebra is introduced and used throughout the last half of this chapter. The study of compact operators and Fredholm operators is taken up in the fifth chapter along with certain ancillary results concerning ideals in \( {C}^{ * } \) -algebras. The approach here is a bit unorthodox but is suggested by modern developments. The last two chapters are of a slightly different character and present a systematic development including recent research of the theory of Toeplitz operators. This latter class of operators has attracted the attention of several mathematicians recently and occurs in several rather diverse contexts. In the sixth chapter certain topics from the theory of Hardy spaces are developed. The selection is dictated by needs of the last chapter and proofs are based on the techniques obtained earlier in the book. The study of Toeplitz operators is taken up in the seventh chapter. Most of what is known in the scalar case is presented including Widom's result on the connectedness of the spectrum. At the end of each chapter there are source notes which suggest additional reading along with giving some comments on who proved what and when. Although a reasonable attempt has been made in the latter chapters at citing the appropriate source for important results, omissions have undoubtedly occurred. Moreover, the absence of a reference should not be construed to mean the result is due to the author. In addition, following each chapter is a large number of problems of varying difficulty. The purposes of these are many: to allow the reader to test his understanding; to indicate certain extensions of the theory which are now accessible; to alert the reader to certain important and related results of which he should be aware along with a hint or a reference for the proof; and to point out certain questions for which the answer is not known. These latter questions are indicated by a double asterisk; a single asterisk indicates a difficult problem. August 1971 ## Acknowledgments This book began as a set of lecture notes for a course given at the University of Michigan in Spri
1013_(GTM179)Banach Algebra Techniques in Operator Theory
1
giving some comments on who proved what and when. Although a reasonable attempt has been made in the latter chapters at citing the appropriate source for important results, omissions have undoubtedly occurred. Moreover, the absence of a reference should not be construed to mean the result is due to the author. In addition, following each chapter is a large number of problems of varying difficulty. The purposes of these are many: to allow the reader to test his understanding; to indicate certain extensions of the theory which are now accessible; to alert the reader to certain important and related results of which he should be aware along with a hint or a reference for the proof; and to point out certain questions for which the answer is not known. These latter questions are indicated by a double asterisk; a single asterisk indicates a difficult problem. August 1971 ## Acknowledgments This book began as a set of lecture notes for a course given at the University of Michigan in Spring, 1968 and again at SUNY at Stony Brook in the academic year, 1969-1970. I am indebted to many people in the writing of this book. Most of all I would like to thank Bruce Abrahamse who prepared the original notes and who has been a constant source of suggestions and constructive criticism writing portions of later versions. In addition, I would like to thank many friends and colleagues for their many suggestions and, in particular, Paul Halmos, Carl Pearcy, Pasquale Porcelli, Donald Sarason, and Allen Shields with whom I have learned many of the things presented in this book. Special thanks are due to Berrien Moore III who read and criticized the entire manuscript and to Joyce Lemen, Dorothy Lentz, and Carole Alberghine for typing the various versions of this manuscript. Lastly, I would like to thank the National Science Foundation and the Alfred E. Sloan Foundation for various support during the writing of this book. August 1971 ## Contents Preface to the Second Edition v Preface to the First Edition vii Acknowledgments ix Symbols and Notation XV 1 Banach Spaces 1 1 The Banach Space of Continuous Functions 2 2 Abstract Banach Spaces . 2 3 The Conjugate Space of Continuous Linear Functionals 5 4 Examples of Banach spaces: \( {c}_{0},{l}^{1} \), and \( {l}^{\infty } \) 6 5 Weak Topologies on Banach Spaces 8 6 The Alaoglu Theorem .9 7 The Hahn-Banach Theorem 10 8 The Conjugate Space of \( C\left( \left\lbrack {0,1}\right\rbrack \right) \) 12 9 The Open Mapping Theorem 21 10 The Lebesgue Spaces: \( {L}^{1} \) and \( {L}^{\infty } \) 23 11 The Hardy Spaces: \( {H}^{1} \) and \( {H}^{\infty } \) 25 Notes 26 Exercises 26 2 Banach Algebras 30 1 The Banach Algebra of Continuous Functions 30 2 Abstract Banach Algebras 31 3 Abstract Index in a Banach Algebra 33 4 The Space of Multiplicative Linear Functions . 36 5 The Gelfand Transform 37 6 The Gelfand-Mazur Theorem 39 7 The Gelfand Theorem for Commutative Banach Algebras 41 8 The Spectral Radius Formula. 42 9 The Stone-Weierstrass Theorem 43 10 The Generalized Stone-Weierstrass Theorem 43 11 The Disk Algebra . 47 12 The Algebra of Functions with Absolutely Convergent Fourier Series .50 13 The Algebra of Bounded Measurable Functions 52 Notes 53 Exercises 53 3 Geometry of Hilbert Space 58 1 Inner Product Spaces 58 2 The Cauchy-Schwarz Inequality 59 3 The Pythagorean Theorem . 60 4 Hilbert Spaces 61 5 Examples of Hilbert Spaces: \( {\mathbb{C}}^{n},{l}^{2},{L}^{2} \), and \( {H}^{2} \) 61 6 The Riesz Representation Theorem 66 7 The Existence of Orthonormal Bases 69 8 The Dimension of Hilbert Spaces 70 Notes 71 Exercises 71 4 Operators on Hilbert Space and \( {C}^{ * } \) -Algebras 74 1 The Adjoint Operator . 75 2 Normal and Self-adjoint Operators 77 3 Projections and Subspaces 78 4 Multiplication Operators and Maximal Abelian Algebras. 80 5 The Bilateral Shift Operator. 82 \( 6{C}^{ * } \) -Algebras 83 7 The Gelfand-Naimark Theorem 84 8 The Spectral Theorem 85 9 The Functional Calculus 85 10 The Square Root of Positive Operators 86 11 The Unilateral Shift Operator 87 12 The Polar Decomposition 88 13 Weak and Strong Operator Topologies 91 14 \( {W}^{ * } \) -Algebras 92 15 Isomorphisms of \( {L}^{\infty } \) -Spaces 94 16 Normal Operators with Cyclic Vectors 95 17 Maximal Abelian \( {W}^{ * } \) -Algebras 97 18 *-Homomorphisms of \( {C}^{ * } \) -Algebras 100 19 The Extended Functional Calculus 101 20 The Fuglede Theorem 103 Notes 104 Exercises 104 5 Compact Operators, Fredholm Operators, and Index Theory 108 1 The Ideals of Finite Rank and Compact Operators 108 2 Approximation of Compact Operators. 110 3 Examples of Compact Operators: Integral Operators 112 4 The Calkin Algebra and Fredholm Operators 113 5 Atkinson's Theorem. 114 6 The Index of Fredholm Operators 115 7 The Fredholm Alternative 116 8 Volterra Integral Operators 118 9 Connectedness of the Unitary Group in a \( {W}^{ * } \) -Algebra 119 10 Characterization of Index . 123 11 Quotient \( {C}^{ * } \) -Algebras 124 12 Representations of the \( {C}^{ * } \) -Algebra of Compact Operators 126 Notes 129 Exercises 129 6 The Hardy Spaces 133 1 The Hardy Spaces: \( {H}^{1},{H}^{2} \), and \( {H}^{\infty } \) 133 2 Reducing Subspaces of Unitary Operators 135 3 Beurling's Theorem 136 4 The F. and M. Riesz Theorem 137 5 The Maximal Ideal Space of \( {H}^{\infty } \) . 138 6 The Inner-Outer Factorization of Functions in \( {H}^{2} \) 141 7 The Modulus of Outer Functions . 141 8 The Conjugates of \( {H}^{1} \) and \( {L}^{\infty }/{H}_{0}^{\infty } \) 144 9 The Closedness of \( {H}^{\infty } + C \) . 145 10 Approximation by Quotients of Inner Functions 145 11 The Gleason-Whitney Theorem 146 12 Subalgebras between \( {H}^{\infty } \) and \( {L}^{\infty } \) 147 13 Abstract Harmonic Extensions 149 14 The Maximal Ideal Space of \( {H}^{\infty } + C \) 150 15 The Invertibility of Functions in \( {H}^{\infty } + C \) 152 Notes 153 Exercises 153 7 Toeplitz Operators 158 1 Toeplitz Operators 158 2 The Spectral Inclusion Theorem 160 3 The Symbol Map 160 4 The Spectrum of Self-adjoint Toeplitz Operators 163 5 The Spectrum of Analytic Toeplitz Operators 164 6 The \( {C}^{ * } \) -Algebra Generated by the Unilateral Shift 164 7 The Invertibility of Toeplitz Operators with Continuous Symbol . 166 8 The Invertibility of Unimodular Toeplitz Operators and Prediction Theory . . 167 9 The Spectrum of Toeplitz Operators with Symbol in \( {H}^{\infty } + C \) 169 10 The Connectedness of the Essential Spectrum 174 11 Localization to the Center of a \( {C}^{ * } \) -Algebra 177 12 Locality of Fredholmness for Toeplitz Operators 178 Notes 178 Exercises 181 References 185 Index 191 ## Symbols and Notation <table><tr><td>A</td><td>47</td><td>\( {H}_{0}^{p} \)</td><td>144</td></tr><tr><td>Aut( )</td><td>55</td><td>\( {H}^{\infty } \)</td><td>25</td></tr><tr><td></td><td></td><td>\( {H}_{0}^{\infty } \)</td><td>26</td></tr><tr><td>C</td><td>1</td><td>\( {H}^{\infty } + C\left( \mathbb{T}\right) \)</td><td>145</td></tr><tr><td>\( {\mathbb{C}}^{n} \)</td><td>61</td><td></td><td></td></tr><tr><td>\( C\left( X\right) \)</td><td>1</td><td>\( {I}_{E} \)</td><td>17</td></tr><tr><td>\( {\chi }_{n} \)</td><td>25</td><td>\( {i}_{t} \)</td><td>166</td></tr><tr><td>\( {c}_{0}\left( {\mathbb{Z}}^{ + }\right) \)</td><td>6</td><td></td><td></td></tr><tr><td>\( {\mathfrak{C}}_{T} \)</td><td>85</td><td>\( j \)</td><td>115</td></tr><tr><td>D</td><td>48</td><td>ker</td><td>75</td></tr><tr><td>\( \partial {F}_{\lambda } \)</td><td>177</td><td></td><td></td></tr><tr><td></td><td></td><td>\( {l}^{1}\left( {\mathbb{Z}}^{ + }\right) \)</td><td>6</td></tr><tr><td>exp</td><td>34</td><td>\( {l}^{1}\left( \mathbb{Z}\right) \)</td><td>50</td></tr><tr><td></td><td></td><td>\( {l}^{2}\left( {\mathbb{Z}}^{ + }\right) \)</td><td>62</td></tr><tr><td>F( )</td><td>113</td><td>\( {l}^{2}\left( \mathbb{Z}\right) \)</td><td>82</td></tr><tr><td>\( {\mathfrak{F}}_{n} \)</td><td>115</td><td>\( {l}^{\infty }\left( {\mathbb{Z}}^{ + }\right) \)</td><td>6</td></tr><tr><td>\( {F}_{\lambda } \)</td><td>177</td><td>\( {L}^{1} \)</td><td>23</td></tr><tr><td></td><td></td><td>\( {L}^{2} \)</td><td>63</td></tr><tr><td>S</td><td>33</td><td>\( {L}^{p} \)</td><td>23</td></tr><tr><td>F93</td><td>37</td><td>\( {L}^{\infty } \)</td><td>23</td></tr><tr><td></td><td></td><td>\( \mathfrak{L}\left( \;\right) \)</td><td>20</td></tr><tr><td>\( h \)</td><td>163</td><td>\( \mathfrak{{LC}}\left( \text{ 全 }\right) \)</td><td>108</td></tr><tr><td>\( {\mathbf{H}}^{1} \)</td><td>25</td><td>\( \mathfrak{{LF}}\left( \text{ }\right) \)</td><td>108</td></tr><tr><td>\( {H}_{0}^{1} \)</td><td>144</td><td>A93</td><td>33</td></tr><tr><td>\( {H}^{2} \)</td><td>64</td><td></td><td></td></tr><tr><td>\( {H}^{p} \)</td><td>133</td><td>\( {M}_{93} \)</td><td>36</td></tr></table> <table><tr><td colspan="4">AVI SynDODS and Notation</td></tr><tr><td>\( {M}_{\varphi } \)</td><td>81</td><td>\( {T}_{\varphi } \)</td><td>158</td></tr><tr><td>\( {M}_{\infty } \)</td><td>138</td><td>\( {T}_{K} \)</td><td>112</td></tr><tr><td>\( M\left( X\right) \)</td><td>19</td><td>\( \mathfrak{T}\left( \text{ }\right) \)</td><td>160</td></tr><tr><td>\( P \)</td><td>158</td><td>\( U \)</td><td>82</td></tr><tr><td>P</td><td>47</td><td>\( {U}_{ + } \)</td><td>87</td></tr><tr><td>\( {\mathcal{P}}_{ + } \)</td><td>47</td><td>\( {U}_{ + }^{\left( n\right) } \)</td><td>121</td></tr><tr><td>PC</td><td>157</td><td></td><td></td></tr><tr><td>\( {P}_{\mathcal{M}} \)</td><td>78</td><td>\( w\left( \text{ }\right) \) \( W\left( \text{ 等 }\right) \)</td><td>104 104</td></tr><tr><td>\( {QC} \)</td><td>157</td><td>\( {\mathfrak{W}}_{T} \)</td><td>92</td></tr><tr><td>R</td><td>3</td><td>Z</td><td>6</td></tr><tr><td>\( r\left( \right) \)</td><td>38</td><td></td><td></td></tr><tr><td>R( )</td><td>52</td><td>\( \begin{array}{l} \parallel \\ * \\ * \end{array}
1013_(GTM179)Banach Algebra Techniques in Operator Theory
2
</tr><tr><td>\( {M}_{\varphi } \)</td><td>81</td><td>\( {T}_{\varphi } \)</td><td>158</td></tr><tr><td>\( {M}_{\infty } \)</td><td>138</td><td>\( {T}_{K} \)</td><td>112</td></tr><tr><td>\( M\left( X\right) \)</td><td>19</td><td>\( \mathfrak{T}\left( \text{ }\right) \)</td><td>160</td></tr><tr><td>\( P \)</td><td>158</td><td>\( U \)</td><td>82</td></tr><tr><td>P</td><td>47</td><td>\( {U}_{ + } \)</td><td>87</td></tr><tr><td>\( {\mathcal{P}}_{ + } \)</td><td>47</td><td>\( {U}_{ + }^{\left( n\right) } \)</td><td>121</td></tr><tr><td>PC</td><td>157</td><td></td><td></td></tr><tr><td>\( {P}_{\mathcal{M}} \)</td><td>78</td><td>\( w\left( \text{ }\right) \) \( W\left( \text{ 等 }\right) \)</td><td>104 104</td></tr><tr><td>\( {QC} \)</td><td>157</td><td>\( {\mathfrak{W}}_{T} \)</td><td>92</td></tr><tr><td>R</td><td>3</td><td>Z</td><td>6</td></tr><tr><td>\( r\left( \right) \)</td><td>38</td><td></td><td></td></tr><tr><td>R( )</td><td>52</td><td>\( \begin{array}{l} \parallel \\ * \\ * \end{array} \)</td><td>1</td></tr><tr><td>ran</td><td>75</td><td></td><td>5, 75, 107</td></tr><tr><td>\( \rho \left( \;\right) \)</td><td>38</td><td>( ),</td><td>9</td></tr><tr><td>\( {\rho }_{\mathrm{e}}\left( \;\right) \)</td><td>171</td><td>\( \parallel \;{\parallel }_{p} \)</td><td>23</td></tr><tr><td></td><td></td><td>田</td><td>27, 72</td></tr><tr><td>\( \sigma \left( \text{ }\right) \)</td><td>38</td><td>\( \bot \)</td><td>28, 65</td></tr><tr><td>\( \sum \)</td><td>3</td><td>\( \otimes \)</td><td>29, 56, 73, 106</td></tr><tr><td></td><td></td><td>(,)</td><td>58</td></tr><tr><td>T</td><td>25</td><td>1</td><td>98</td></tr></table> ## Banach Spaces 1.1 We begin by introducing the most representative example of a Banach space. Let \( X \) be a compact Hausdorff space and let \( C\left( X\right) \) denote the set of continuous complex-valued functions on \( X \) . For \( {f}_{1} \) and \( {f}_{2} \) in \( C\left( X\right) \) and \( \lambda \) a complex number, we define: (1) \( \left( {{f}_{1} + {f}_{2}}\right) \left( x\right) = {f}_{1}\left( x\right) + {f}_{2}\left( x\right) \) ; (2) \( \left( {\lambda {f}_{1}}\right) \left( x\right) = \lambda {f}_{1}\left( x\right) \) ; and (3) \( \left( {{f}_{1}{f}_{2}}\right) \left( x\right) = {f}_{1}\left( x\right) {f}_{2}\left( x\right) \) . With these operations \( C\left( X\right) \) is a commutative algebra with identity over the complex field \( \mathbb{C} \) . Each function \( f \) in \( C\left( X\right) \) is bounded, since it follows from the fact that \( f \) is continuous and \( X \) is compact that the range of \( f \) is a compact subset of \( \mathbb{C} \) . Thus the least upper bound of \( \left| f\right| \) is finite; we call this number the norm of \( f \) and denote it by \[ \parallel f{\parallel }_{\infty } = \sup \{ \left| {f\left( x\right) }\right| : x \in X\} . \] The following properties of the norm are easily verified: (1) \( \parallel f{\parallel }_{\infty } = 0 \) if and only if \( f = 0 \) ; (2) \( \parallel {\lambda f}{\parallel }_{\infty } = \left| \lambda \right| \parallel f{\parallel }_{\infty } \) ; (3) \( \parallel f + g{\parallel }_{\infty } \leq \parallel f{\parallel }_{\infty } + \parallel g{\parallel }_{\infty } \) ; and (4) \( \parallel {fg}{\parallel }_{\infty } \leq \parallel f{\parallel }_{\infty }\parallel g{\parallel }_{\infty } \) . We define a metric \( \rho \) on \( C\left( X\right) \) by \( \rho \left( {f, g}\right) = \parallel f - g{\parallel }_{\infty } \) . The properties of a metric, namely, (1) \( \rho \left( {f, g}\right) = 0 \) if and only if \( f = g \) , (2) \( \rho \left( {f, g}\right) = \rho \left( {g, f}\right) \), and (3) \( \rho \left( {f, h}\right) \leq \rho \left( {f, g}\right) + \rho \left( {g, h}\right) \) , follow immediately from properties (1)-(3) of the norm. It is easily seen that convergence with respect to the metric \( \rho \) is just uniform convergence. An important property of this metric is that \( C\left( X\right) \) is complete with respect to it. 1.2 Proposition. If \( X \) is a compact Hausdorff space, then \( C\left( X\right) \) is a complete metric space. Proof If \( {\left\{ {f}_{n}\right\} }_{n = 1}^{\infty } \) is a Cauchy sequence, then \[ \left| {{f}_{n}\left( x\right) - {f}_{m}\left( x\right) }\right| \leq {\begin{Vmatrix}{f}_{n} - {f}_{m}\end{Vmatrix}}_{\infty } = \rho \left( {{f}_{n},{f}_{m}}\right) \] for each \( x \) in \( X \) . Hence, \( {\left\{ {f}_{n}\left( x\right) \right\} }_{n = 1}^{\infty } \) is a Cauchy sequence of complex numbers for each \( x \) in \( X \), so we may define \( f\left( x\right) = \mathop{\lim }\limits_{{n \rightarrow \infty }}{f}_{n}\left( x\right) \) . We need to show that \( f \) is in \( C\left( X\right) \) and that \( \mathop{\lim }\limits_{{n \rightarrow \infty }}{\begin{Vmatrix}f - {f}_{n}\end{Vmatrix}}_{\infty } = 0 \) . To that end, given \( \varepsilon > O \), choose \( N \) such that \( n, m \geq N \) implies \( {\begin{Vmatrix}{f}_{n} - {f}_{m}\end{Vmatrix}}_{\infty } < \varepsilon \) . For \( {x}_{0} \) in \( X \) there exists a neighborhood \( U \) of \( {x}_{0} \) such that \( \left| {{f}_{N}\left( {x}_{0}\right) - {f}_{N}\left( x\right) }\right| < \varepsilon \) for \( x \) in \( U \) . Therefore, \[ \left| {f\left( {x}_{0}\right) - f\left( x\right) }\right| \leq \mathop{\lim }\limits_{{n \rightarrow \infty }}\left| {{f}_{n}\left( {x}_{0}\right) - {f}_{N}\left( {x}_{0}\right) }\right| + \mathop{\lim }\limits_{{n \rightarrow \infty }}\left| {{f}_{N}\left( {x}_{0}\right) - {f}_{N}\left( x\right) }\right| \] \[ + \mathop{\lim }\limits_{{n \rightarrow \infty }}\left| {{f}_{N}\left( x\right) - {f}_{n}\left( x\right) }\right| \] \[ \leq {3\varepsilon } \] which implies \( f \) is continuous. Further, for \( n \geq N \) and \( x \) in \( X \), we have \[ \left| {{f}_{n}\left( x\right) - f\left( x\right) }\right| = \left| {{f}_{n}\left( x\right) - \mathop{\lim }\limits_{{m \rightarrow \infty }}{f}_{m}\left( x\right) }\right| = \mathop{\lim }\limits_{{m \rightarrow \infty }}\left| {{f}_{n}\left( x\right) - {f}_{m}\left( x\right) }\right| \] \[ \leq \mathop{\limsup }\limits_{{m \rightarrow \infty }}{\begin{Vmatrix}{f}_{n} - {f}_{m}\end{Vmatrix}}_{\infty } \leq \varepsilon \] Thus \( \mathop{\lim }\limits_{{n \rightarrow \infty }}{\begin{Vmatrix}{f}_{x} - f\end{Vmatrix}}_{\infty } = 0 \) and hence \( C\left( X\right) \) is complete. We next define the notion of Banach space which abstracts the salient properties of the preceding example. We shall see later in this chapter that every Banach space is isomorphic to a subspace of some \( C\left( X\right) \) . 1.3 Definition. A Banach space is a complex linear space \( \mathcal{X} \) with a norm \( \parallel \parallel \) satisfying (1) \( \parallel f\parallel = 0 \) if and only if \( f = 0 \) , (2) \( \parallel {\lambda f}\parallel = \left| \lambda \right| \parallel f\parallel \) for \( \lambda \) in \( \mathbb{C} \) and \( f \) in \( \mathcal{X} \), and (3) \( \parallel f + g\parallel \leq \parallel f\parallel + \parallel g\parallel \) for \( f \) and \( g \) in \( \mathcal{X} \) , such that \( \mathcal{X} \) is complete in the metric given by this norm. 1.4 Proposition. Let \( \mathcal{X} \) be a Banach space. The functions \( a : \mathcal{X} \times \mathcal{X} \rightarrow \mathcal{X} \) defined \( a\left( {f, g}\right) = f + g \) , \( s : \mathbb{C} \times \mathcal{X} \rightarrow \mathcal{X} \) defined \( s\left( {\lambda, f}\right) = {\lambda f} \), and \( n : X \rightarrow {\mathbb{R}}^{ + } \) defined \( n\left( f\right) = \parallel f\parallel \) are continuous. Proof Obvious. 1.5 Directed Sets and Nets The topology of a metric space can be described in terms of the sequences in it that converge. For more general topological spaces a notion of generalized sequence is necessary. In what follows it will often be convenient to describe a topology in terms of its convergent generalized sequences. Thus we proceed to review for the reader the notion of net. A directed set \( A \) is a partially ordered set having the property that for each pair \( \alpha \) and \( \beta \) in \( A \) there exists \( \gamma \) in \( A \) such that \( \gamma \geq \alpha \) and \( \gamma \geq \beta \) . A net is a function \( \alpha \rightarrow {\lambda }_{\alpha } \) on a directed set. If the \( {\lambda }_{\alpha } \) all lie in a topological space \( X \), then the net is said to converge to \( \lambda \) in \( X \) if for each neighborhood \( U \) of \( \lambda \) there exists \( {\alpha }_{U} \) in \( A \) such that \( {\lambda }_{\alpha } \) is in \( U \) for \( \alpha \geq {\alpha }_{U} \) . Two topologies on a space \( X \) coincide if they have the same convergent nets. Lastly, a topology can be defined on \( X \) by prescribing the convergent nets. For further information concerning nets and subnets, the reader should consult [71]. We now consider the convergence of Cauchy nets in a Banach space. 1.6 Definition. A net \( {\left\{ {f}_{\alpha }\right\} }_{\alpha \in A} \) in a Banach space \( X \) is said to be a Cauchy net if for every \( \varepsilon > 0 \), there exists \( {\alpha }_{0} \) in \( A \) such that \( {\alpha }_{1},{\alpha }_{2} \geq {\alpha }_{0} \) implies \( \begin{Vmatrix}{{f}_{{\alpha }_{1}} - {f}_{{\alpha }_{2}}}\end{Vmatrix} < \varepsilon \) . ## 1.7 Proposition. In a Banach space each Cauchy net is convergent. Proof Let \( {\left\{ {f}_{\alpha }\right\} }_{\alpha \in A} \) be a Cauchy net in the Banach space \( \mathcal{X} \) . Choose \( {\alpha }_{1} \) such that \( \alpha \geq {\alpha }_{1} \) implies \( \begin{Vmatrix}{{f}_{\alpha } - {f}_{{\alpha }_{1}}}\end{Vmatrix} < 1 \) . Having chosen \( {\left\{ {\alpha }_{k}\right\} }_{k = 1}^{n} \) in \( A \), choose \( {\alpha }_{{n}_{ + }1} \geq {\alpha }_{n} \) such that \( \alpha \geq {\alpha }_{n + 1} \) implies \[ \begin{Vmatrix}{{f}_{\alpha } - {f}_{{\alpha }_{n + 1}}}\end{Vmatrix} < \frac{1}{n + 1} \] The sequence \( {