Thursday, April 22, 2021

Growth of the Gaussian pivoting algorithm

Gaussian elimination” is the standard method for solving systems of linear equations that runs by choosing one pivot variable in one of the equations and eliminating it from the other equations by a suitable linear combination. These other equations have one less variable and one may iterate the process, leading to a system that has a triangular form and can be solved in a relatively straightforward way. In the so-called Gauss-Jordan method, the pivot variables are eliminated from all of the equations, and this leads to the row reduced echelon form of the initial system, a new system in which the pivot variables are explicitly solved in terms of the remaining variables; it also has the merit of being independent of the intermediate steps, but this would be a story for another night.

How the name of Gauss is attributed to this method is also another story, that is recounted in great detail by J. Grcar. As he explains, systems of linear equations and their solution already appear on Babylonian tablets (2000 BC), on the Egyptian Rhind papyrus (1550 BC), in the Chinese Nine chapters on the mathematical art (200 AD), the Arithmetica of the Greek mathematician Diophantus (250 AD), the Āryabhaṭīya of the Hindu mathematician Āryabhaṭa (499 AD). Such systems were essentially absent of the mathematics treatises during the Renaissance and it is to Newton (17th century) that we owe its revival in Western Europe. At the beginning of the 19th century, in relation with the problem in celestial mechanics/geodesy of fitting together multiple imprecise measurements, Gauss and Legendre invented the least square methods. This involved a system of linear equations which Gauss solved by what he called “common elimination”. In the 20th century, the development of computing machines and numerical analysis led to further work, from Cholesky (in relation with geodesy), Crout, to von Neumann and Goldstine and their LDULDU decomposition.

Whoever had to perform elimination by hand knows that the computations are rapidly tedious and often lead to more and more complicated fractions. 

When computer calculations are done with floating point algebra, the difficulty of rounding errors appears. If in a linear system, say Ax=bAx=b,  the matrices AA and bb are only known up to an error, so that the system that is actually solved would rather be (A+E)x=b+δb(A+E)x=b+\delta b, and it is of an obvious importance to compare its solution with the solution of the initial system. One way to make this comparison involves the inverse of AA, which is unknown at this stage. The product of the norms AA1\|A\| \,\|A^{-1}\| is the conditioning number of AA, and one can not avoid the problem of ill-conditioned matrices, which will inherently lead to lack of precision.

But when floating points numbers are used in the computation, a new problem appears, even when one restricts to well-conditioned systems. Floating points numbers are of the form ±a10e\pm a\cdot 10^e (in base 10, say), where aa is a real number between 11 and 1010 which is known up to a fixed number of decimal places. In other words, floating points numbers are known up to a relative error. Let us examine the consequence for the subtraction of floating point numbers. Consider two such numbers, say x=a10ex=a\cdot 10^e and x=a10ex'=a'\cdot 10^e, with the same exponent ee, and such that aa and aa' are close. Their difference is given by xx=(aa)10ex'-x=(a'-a)\cdot 10^e, but since aa and aa' are close, their difference y=xxy=x'-x is no more between 11 and 1010, but may be, say of size 10510^{-5}, so that the floating point expression of xxx'-x is (105(aa))10e5=b10e5(10^5(a'-a))\cdot 10^{e-5}=b\cdot 10^{e-5}; the problem is that the last 5 decimals of bb are absolutely unknown, which leads to a relative error for yy which is 10510^5 as big as expected!

Now, by its very essence, the elimination method features a lot of such subtractions, hence it is inherently not very well suited with floating points numbers. 

In the 1960s, the American mathematician Wilkinson analysed the situation very precisely.  He showed that for the Gaussian elimination, the main parameter is the relative size of the matrices that intervene in the process. To set up some notation, imagine that the initial system Ax=bAx=b is transformed, step by step, into a series of equivalent systems A(r)x=b(r)A^{(r)}x=b^{(r)}, where A(r)=(aij(r))A^{(r)}=(a_{ij}^{(r)}) is a matrix whose first rr lines are in triangular form, and the remaining nrn-r lines still need to be worked on. To get the matrix A(r+1)A^{(r+1)}, one subtract a multiple mirm_{ir} of the rrth row from the iith row he multipliers, for ii ranging from r+1r+1 to nn, where the multipliers mirm_{ir} are defined by
mir=air(r)/arr(r). m_{ir}=a_{ir}^{(r)}/a_{rr}^{(r)}.
Large multipliers lead to lack of precision, but if complete pivoting method is used, one has mir1|m_{ir}|\leq 1. In this case, one observes that a bound aij(r)M|a_{ij}^{(r)}|\leq M for the coefficients of A(r)A^{(r)} leads to the bound aij(r+1)2M|a_{ij}^{(r+1)}|\leq 2M at the next step. At the last step, one gets aij(n)2n1M|a_{ij}^{(n)}|\leq 2^{n-1}M, where M=sup(aij)M=\sup(|a_{ij}|). Consequently, the relevant constant to be estimated,
R(A)=suprsupi,jaij(r)supi,jaij, R(A) = \sup_r \frac{\sup_{i,j}|a_{ij}^{(r)}|}{\sup_{i,j}|a_{ij}|},
satisfies R(A)2n1R(A)\leq 2^{n-1}.

In fact, Wilkinson (1961) gave a much better bound. Let B(r)B^{(r)} be the square matrix of size nrn-r that has to be worked on after the rrth step and let brb_{r} be the maximum size of its coefficients, the size of the chosen pivot since one does complete pivoting. One has the following fomula for its determinant:
det(B(r))=br+1bn. \det(B^{(r)})= b_{r+1}\cdots b_n.
Moreover, the Euclidean norms of the the columns of B(r)B^{(r)} are bounded above by nrbr+1\sqrt{n-r} b_{r+1} and Hadamard inequality (“the volume of a parallelepiped is smaller than the product of the sizes of its edges”) implies that
det(B(r))(nr)(nr)/2br+1nr. | \det(B^{(r)})| \leq (n-r)^{(n-r)/2} b_{r+1}^{n-r}.
Together, these relations lead to an upper bound
R(A)n1/2(231/241/3n1/(n1))1/2, R(A) \leq n^{1/2} \left( 2\cdot 3^{1/2}\cdot 4^{1/3}\cdots n^{1/(n-1)}\right)^{1/2},
roughly nlog(n)/4n^{\log(n)/4}, a bound that Wilkinson considers to be a “severe overestimate”, and in Wilkinson (Rounding Errors in Algebraic Processes, Dover, 1963, bottom of p. 97), he even notes that “No matrix has yet been discovered for which R(A)>nR(A)>n.

This statement remained known as Wilkinson's conjecture, although Wilkinson himself did not state it as such. Tornheim (1964) proved that Hadamard matrices — matrices with entries ±1\pm1 and pairwise orthogonal columns and rows — satisfy R(A)nR(A)\geq n and this led Cryer (1968) to formally state the conjecture and to suspect that R(A)<nR(A)<n unless AA is a Hadamard matrix. Some matrices have been shown such that R(A)(n+1)/2R(A)\geq (n+1)/2 (Higham & Higham, 1989) and random matrices seems to have an RR-factor roughly n1/2n^{1/2} (Trefethen & Schreiber, 1990). In fact, Hadamard matrices of size n=12n=12 (Edelman & Mascarenhas, 1995) or n=16n=16 (Kravvaritis & Mtrouli, 2009) satisfy R(A)=nR(A)=n, but this is definitely nontrivial.

However, Gould (1991) could exhibit a matrix AA of size n=13n=13 with  growth factor R(A)=13.0205R(A)=13.0205, thus providing a counterexample to Wilkinson's “conjecture”. To that aim, he reduced the question to finding a solution of a nonlinear programming problem with roughly n3/3700n^3/3\approx 700 variables, fortunately a sparse one, a computation he did using a programming package he had developed with Conn and Toint. The matrix he gives has integer coefficients of size up to 20 digits!

But the story does not end here!

Edelman tried to replicate Gould's computations using the computer algebra softwares Mathematica and Maple — what he found is the growth factor of Gould's matrix is around 7.3557.355, consistently in both softwares! As he writes, one of these softwares could have had a bug, but it is rather unlikely that both of them had had the same one.

What happened, and you will agree that there is much irony in this story, is that Gould had performed his computations using floating point algebra. A near tie in the 6th pivot lead to an incorrect choice of pivot and to an erroneous computation, even within the double precision that he has used.

Fortunately, Edelman (1992) showed that changing one coefficient by 10710^{-7} in Gould's matrix yields a growth of 13.0213.02, so that Wilkinson's “conjecture” is definitely incorrect.

Friday, April 2, 2021

On the Hadamard-Lévy theorem, or is it Banach-Mazur?

During the preparation of an agrégation lecture on connectedness, I came across the following theorem, attributed to Hadamard–Lévy: 

Theorem. — Let f ⁣:RnRnf\colon \mathbf R^n\to\mathbf R^n be a C1\mathscr C^1-map which is proper and a local diffeomorphism. Then ff is a global diffeomorphism.

In this context, that ff is proper means that f(x)+\| f(x)\| \to+\infty when x+\| x\|\to+\infty, while, by the inverse function theorem, the condition that ff is a local diffeomorphism is equivalent to the property that its differential f(x)f'(x) is invertible, for every xRnx\in\mathbf R^n. The conclusion is that ff is a diffeomorphism from Rn\mathbf R^n to itself; in particular, ff is bijective and its inverse is continuous.

This theorem is not stated in this form neither by Hadamard (1906), nor by Lévy (1920), but is essentially due to Banach & Mazur (1934) and it is the purpose of this note to clarify the history, explain a few proofs, as well as more recent consequences for partial differential equations.

A proper map is closed: the image f(A)f(A) of a closed subset AA of Rn\mathbf R^n is closed in Rn\mathbf R^n. Indeed, let (am)(a_m) be a sequence in AA whose image (f(am))(f(a_m)) converges in Rn\mathbf R^n to an element bb; let us show that there exists aAa\in A such that b=f(a)b=f(a). The properness assumption on ff implies that (am)(a_m) is bounded. Consequently, it has a limit point aa, and aAa\in A because AA is closed. Necessarily, f(a)f(a) is a limit point of the sequence (f(am))(f(a_m)), hence b=f(a)b=f(a).

In this respect, let us note the following reinforcement of the previous theorem, due to Browder (1954):
Theorem (Browder). — Let f ⁣:RnRnf\colon \mathbf R^n\to\mathbf R^n be a local homeomorphism. If ff is closed, then ff is a global homeomorphism.

A surprising aspect of these results and their descendents is that they are based on two really different ideas. Banach & Mazur and Browder are based on the notion of covering, with ideas of homotopy theory and, ultimately, the fact that Rn\mathbf R^n is simply connected. On the other hand, the motivation of Hadamard was to generalize to dimension nn the following elementary discussion in the one-dimensional case: Let f ⁣:RRf\colon\mathbf R\to\mathbf R be a C1\mathscr C^1-function whose derivative is >0>0 everywhere (so that ff is strictly increasing); give a condition for ff to be surjective. In this case, the condition is easy to find: the indefinite integral f(x)dx\int f'(x)\,dx has to be divergent both at -\infty and ++\infty. In the nn-dimensional case, the theorems of Hadamard is the following:

Theorem.Let f ⁣:RnRnf\colon\mathbf R^n\to\mathbf R^n be a C1\mathscr C^1-map. For rR+r\in\mathbf R_+, let ω(r)\omega(r) be the infimum, for xRnx\in\mathbf R^n such that x=r\|x\|=r, of the norm of the linear map f(x)1f'(x)^{-1}; if 0dr/ω(r)=+\int_0^\infty dr/\omega(r)=+\infty, then ff is a global diffeomorphism.

In Hadamard's paper, the quantity ω(r)\omega(r) is described geometrically as the minor axis of the ellipsoid defined by f(x)f'(x), and Hadamard insists that using the volume of this ellipsoid only, essentially given by the determinant of f(x)f'(x), would not suffice to characterize global diffeomorphisms. (Examples are furnished by maps of the form f(x1,x2)=(f1(x1),f2(x2))f(x_1,x_2)=(f_1(x_1),f_2(x_2)). The determinant condition considers f1(x1)f2(x2)f_1'(x_1)f_2'(x_2), while one needs individual conditions on f1(x1)f'_1(x_1) and f2(x2)f'_2(x_2).)

In fact, as explained in Plastock (1974), both versions (closedness hypothesis or quantitative assumptions on the differential) imply that the map ff is a topological covering of Rn\mathbf R^n. Since the target Rn\mathbf R^n is simply connected and the source Rn\mathbf R^n is connceted, ff has to be a homeomorphism. I will explain this proof below, but I would first like to explain another one, due to Zuily & Queffelec (1995) propose an alternating proof which is quite interesting.

A dynamical system approach

The goal is to prove that ff is bijective and, to that aim, we will prove that every preimage set f1(b)f^{-1}(b) is reduced to one element. Replacing ff by fbf-b, it suffices to treat the case of b=0b=0. In other words, we wish to solve that the equation f(x)=0f(x)=0 has exactly one solution. For that, it is natural to try to start from some point ξRn\xi\in\mathbf R^n and to force ff to decrease. This can be done by following the flow of the vector field given by v(x)=f(x)1(f(x))v(x)=-f'(x)^{-1}(f(x)). This is a vector field on Rn\mathbf R^n and we can consider its flow: a map Φ\Phi defined on an open subset of R×Rn\mathbf R\times\mathbf R^n such that tΦ(t,x)=v(Φ(t,x))\partial_t \Phi(t,x)=v(\Phi(t,x)) for all (t,x)(t,x) and Φ(0,x)=x\Phi(0,x)=x for all xx. In fact, the Cauchy–Lipschitz theorem guarantees the existence of such a flow only if the vector field vv is locally Lipschitz, which happens if, for example, ff is assumed to be C2\mathscr C^2. In this case, there is even uniqueness of a maximal flow, and we will make this assumption, for safety. (In fact, the paper of De Marco, Gorni & Zampieri (1994) constructs the flow directly thanks to the hypothesis that the vector field is pulled back from the Euler vector field on Rn\mathbf R^n.)

What are we doing here? Note that in Rn\mathbf R^n, the opposite of the Euler vector field, defined by u(y)=yu(y)=-y, has a very simple solution: the flow lines are straight lines going to 00. The formula above just pulls back this vector field uu via the local diffeomorphism ff, and the flow lines of the vector field vv will just be the ones given by pull back by ff, which will explain the behaviour described below.

In particular, let aRna\in\mathbf R^n be such that f(a)=0f(a)=0 and let UU be a neighborhood of aa such that ff induces a diffeomorphism from UU to a ball around 00. Pulling back the solution of the minus-Euler vector field by ff, we see that once a flow line enters the open set UU, it converges to aa. The goal is now to prove that it will indeed enter such a neighborhood (and, in particular, that such a point aa exists).

We consider a flow line starting from a point xx, that is, ϕ(t)=Φ(t,x)\phi(t)=\Phi(t,x) for all times tt. Let g(t)=f(ϕ(t))g(t)= f(\phi(t)); observe that gg satisfies g(t)=f(ϕ(t))(ϕ(t))=g(t)g'(t)=f'(\phi(t))(\phi'(t))=-g(t), hence g(t)=g(0)etg(t)=g(0)e^{-t}. Assume that the line flow is defined on [0;t1[[0;t_1\mathopen[, with t1<+t_1<+\infty. by what precedes, gg is bounded in the neighborhood of t1t_1; since ff is assumed to be proper, this implies that ϕ(t)\phi(t) is bounded as well. The continuity of the vector field vv implies that ϕ\phi is uniformly continuous, hence it has a limit at t1t_1. We may then extend the line flow a bit right of t1t_1. As a consequence, the line flow is defined for all times, and g(t)0g(t)\to0 when t+t\to+\infty. By the same properness argument, this implies that ϕ(t)\phi(t) is bounded when t+t\to+\infty, hence it has limit points aa which satisfy f(a)=0f(a)=0. Once ϕ\phi enters an appropriate neighborhood of such a point, we have seen that the line flow automatically converges to some point af1(0)a\in f^{-1}(0).

Let us now consider the map λ ⁣:Rnf1(0)\lambda\colon\mathbf R^n\to f^{-1}(0) that associates with a point ξ\xi the limit of the line flow tΦ(t,ξ)t\mapsto \Phi(t,\xi) starting from the initial condition ξ\xi. By continuity of the flow of a vector field depending on the initial condition, the map λ\lambda is continuous. On the other hand, the hypothesis that ff is a local diffeomorphism implies that f1(0)f^{-1}(0) is a closed discrete subset of Rn\mathbf R^n. Since Rn\mathbf R^n is connected, the map λ\lambda is constant. Since one has λ(ξ)=ξ\lambda(\xi)=\xi for every ξf1(0)\xi\in f^{-1}(0), this establishes that f1(0)f^{-1}(0) is reduced to one element, as claimed.

Once ff is shown to be bijective, the fact that it is proper (closed would suffice) implies that its inverse bijection f1f^{-1} is continuous. This concludes the proof.

The theorem of Banach and Mazur

The paper of Banach and Mazur is written in a bigger generality. They consider multivalued continuous maps F ⁣:XYF\colon X\to Y (kk-deutige stetige Abbildungen) by which they mean that for every xx, a subset F(x)F(x) of YY is given, of cardinality kk, the continuity being expressed by sequences: if xnxx_n\to x, one can order, for every nn, the elements of F(xn)={yn,1,,yn,k}F(x_n)=\{y_{n,1},\dots,y_{n,k}\}, as well as the elements of F(x)={y1,,yk}F(x)=\{y_1,\dots,y_k\}, in such a way that yn,jyny_{n,j}\to y_n for all jj. (In their framework, XX and YY are metric spaces, but one could transpose their definition to topological spaces if needed.) They say that such a map is decomposed (zerfällt) if there are continuous functions f1,,fkf_1,\dots,f_k from XX to YY such that F(x)={f1(x),,fk(x)}F(x)=\{f_1(x),\dots,f_k(x)\} for all xXx\in X.

In essence, the definition that Banach and Mazur are proposing contains as a particular case the finite coverings. Namely, if p ⁣:YXp\colon Y\to X is a finite covering of degree kk, then the map xp1(x)x\mapsto p^{-1}(x) is a continuous kk-valued map from XX to YY. Conversely, let us consider the graph ZZ of FF, namely the set of all points (x,y)X×Y(x,y)\in X\times Y such that yF(x)y\in F(x). Then the first projection p ⁣:ZXp\colon Z\to X is a covering map of degree kk, but it is not clear that it has local sections.

It would however not be so surprising to 21st-century mathematicians that if one makes a suitable assumption of simple connectedness on XX, then every such FF should be decomposed. Banach and Mazur assume that XX satisfies two properties:

  1. The space XX is semilocally arcwise connected: for every point xXx\in X and every neighborhood UU of xx, there exists an open neighborhood UU' contained in UU such that for every point xUx'\in U', there exists a path c ⁣:[0;1]Uc\colon[0;1]\to U such that c(0)=xc(0)=x and c(1)=xc(1)=x'. (Semilocally means that the path is not necessarily in UU' but in UU.)
  2. The space XX is arcwise simply connected: two paths c0,c1 ⁣:[0;1]Xc_0,c_1\colon[0;1]\to X with the same endpoints (c0(0)=c1(0)c_0(0)=c_1(0) and c0(1)=c1(1)c_0(1)=c_1(1)) are strictly homotopic — there exists a continuous map h ⁣:[0;1]Xh\colon[0;1]\to X such that h(0,t)=c0(t)h(0,t)=c_0(t) and h(1,t)=c1(t)h(1,t)=c_1(t) for all tt, and h(s,0)=c0(0)h(s,0)=c_0(0) and h(s,1)=c0(1)h(s,1)=c_0(1) for all ss.

Consider a kk-valued continuous map FF from XX to YY, where XX is connected. Banach and Mazur first prove that for every path c ⁣:[0;1]Xc\colon [0;1]\to X and every point y0F(c(0))y_0\in F(c(0)), there exists a continuous function f ⁣:[0;1]Yf\colon[0;1]\to Y such that f(t)F(c(t))f(t)\in F(c(t)) for all tt. To that aim, the consider disjoint neighborhoods V1,,VkV_1,\dots,V_k of the elements of F(c(0))F(c(0)), with y0V1y_0\in V_1, say, and observe that for tt small enough, there is a unique element in F(c(t))V1F(c(t))\cap V_1. This defines a bit of the path cc, and one can go on. Now, given two paths c,cc,c' such that c(0)=c(0)c(0)=c'(0) and c(1)=c(1)c(1)=c'(1), and two maps f,ff,f' as above, they consider a homotopy h ⁣:[0;1]×[0;1]Xh\colon[0;1]\times[0;1]\to X linking cc to cc'. Subdividing this square in small enough subsquares, one see by induction that f(1)=f(1)f(1)=f'(1). (This is analogous to the proof that a topological covering of the square is trivial.) Fixing a point x0Xx_0\in X and a point y0F(x0)y_0\in F(x_0), one gets in this way a map from XX to YY such that F(x)F(x) is equal to f(1)f(1), for every path c ⁣:[0;1]Xc\colon[0;1]\to X such that c(0)=x0c(0)=x_0 and c(1)=xc(1)=x, and every continuous map f ⁣:[0;1]Yf\colon [0;1]\to Y such that f(t)F(c(t))f(t)\in F(c(t)) for all tt and f(0)=y0f(0)=y_0. This furnishes a map from XX to YY, and one proves that it is continuous. If one considers all such maps, for all points in F(x0)F(x_0), one obtains the decomposition of the multivalued map FF.

To prove their version of the Hadamard–Lévy theorem, Banach and Mazur observe that if f ⁣:YXf\colon Y\to X is a local homeomorphism which is proper, then setting F(x)=f1(y)F(x)=f^{-1}(y) gives a multivalued continuous map. It is not obvious that the cardinalities k(x)k(x) of the sets F(x)F(x) are constant, but this follows (if XX is connected) from the fact that ff is both a local homeomorphism and proper. Then FF is decomposed, so that there exist continuous maps g1,,gk ⁣:XYg_1,\dots,g_k\colon X\to Y such that f1(x)={g1(x),,gk(x)}f^{-1}(x)=\{g_1(x),\dots,g_k(x)\} for all xXx\in X. This implies that YY is the disjoint union of the kk connected subsets gj(X)g_j(X). If YY is connected, then ff is a homeomorphism.

The versions of Hadamard and Lévy, after Plastock

Hadamard considered the finite dimensional case, and Lévy extended it to the case of Hilbert spaces.

Plastock considers a Banach-space version of the theorem above: f ⁣:EFf\colon E\to F is a C1\mathscr C^1-map between Banach spaces with invertible differentials and such that, setting ω(r)=infx=rf(x)1\omega(r)=\inf_{\|x\| = r}\|f'(x)^{-1}\|, one has 0ω(r)dr=+\int_0^\infty \omega(r)\,dr=+\infty. Of course, under these hypotheses, the Banach spaces EE and FF are isomorphic, but it may be useful that they are not identical. Note that f(E)f(E) is open in FF, and the proposition that will insure that ff is a global diffeomorphism is the following one, in the spirit of covering theory.

Proposition.(Assuming that ff is a local diffeomorphism.) It suffices to prove that the map ff satisfies the path lifting property: for every point xEx\in E and every C1\mathscr C^1 map c ⁣:[0;1]f(E)c\colon[0;1]\to f(E) such that c(0)=f(x)c(0)=f(x), there exists a C1\mathscr C^1 map d ⁣:[0;1]Ed\colon[0;1]\to E such that c(t)=f(d(t))c(t)=f(d(t)) for all tt and d(0)=cd(0)=c.

The goal is now to prove that ff satisfies this path lifting property. Using that ff is a local homeomorphism, one sees that lifts are unique, and are defined on a maximal subinterval of [0;1][0;1] which is either [0;1][0;1] itself, or of the form [0;s[[0;s\mathclose[. To prevent the latter case, one needs to impose conditions on the norm f(x)1\| f'(x)^{-1}\| such as the one phrased in terms of ω(r)\omega(r) as in the Hadamard–Lévy theorem. In fact, Plastock starts with a simpler case.

Proposition.The path lifting property follows from the following additional hypotheses:

  1. One has f(x)+\|f(x)\|\to+\infty when x+\|x\|\to+\infty;
  2. There exists a positive continuous function M ⁣:R+R+M\colon\mathbf R_+\to\mathbf R_+ such that f(x)1M(x)\|f'(x)^{-1}\|\leq M(\|x\|) for all $x.

Assume indeed that a path cc has a maximal lift dd, defined over the interval [0;s[[0;s\mathclose[. By the hypothesis (i), d(t)d(t) remains bounded when tst\to s, because c(t)=f(d(t))c(t)=f(d(t)) tends to c(s)c(s). Differentiating the relation c(t)=f(d(t))c(t)=f(d(t)), one gets c(t)=f(d(t))(d(t))c'(t)=f'(d(t))(d'(t)), hence d(t)=f(d(t))1(c(t))d'(t)=f'(d(t))^{-1}(c'(t)), so that d(t)M(d(t))c(t)\| d'(t)\|\leq M(\|d(t)\|) \|c'(t)\|. This implies that d\|d'\| is bounded, so that dd is uniformly continuous, hence it has a limit at ss. Then the path dd can be extended by setting d(s)d(s) to this limit and using the local diffeomorphism property to go beyong ss.

The Hadamard–Lévy is related to completeness of some length-spaces. So we shall modify the distance of the Banach space EE as follows: if c ⁣:[0;1]Ec\colon[0;1]\to E is a path in EE, then its length is defined by  (c)=01 f(c(t))11c(t)dt. \ell(c) = \int_0^1 \| f'(c(t))^{-1}\|^{-1} \|{c'(t)}\|\, dt. Observe that f(c(t))11ω(c(t))\|f'(c(t))^{-1}\|^{-1} \geq \omega(\|c(t)\|), so that (c)01ω(c(t))c(t)dt. \ell(c) \geq \int_0^1 \omega(\|c(t)\|) \|{c'(t)}\|\, dt. The modified distance of two points in EE is then redefined as the infimum of the lengths of all paths joining two points.

Lemma.With respect to the modified distance, the space EE is complete.

One proves that (c) c(0)c(1)ω(r)dr\ell(c) \geq \int_{\|{c(0)}\|}^{\|{c(1)}\|}\omega(r)\,dr. Since 0ω(r)dr=+\int_0^\infty \omega(r)\,dr=+\infty, this implies that Cauchy sequences for the modified distance are bounded in EE for the original norm. On the other hand, on any bounded subset of EE, the Banach norm and the modified distance are equivalent, so that they have the same Cauchy sequences.

Other conditions can be derived from Plastock's general theorem. For example, assuming that EE and FF are a Hilbert space HH, he shows that it suffices to assume the existence of a decreasing function λ ⁣:R+R+\lambda\colon\mathbf R_+\to\mathbf R_+ such that f(x)(u),uλ(x)u2\langle f'(x)(u),u\rangle \geq \lambda(\|x\|) \| u\|^2 for all x,yx,y and 0λ(r)dr=+\int_0^\infty \lambda(r)\,dr=+\infty. Indeed, under this assumption, one may set ω(r)=λ(r)\omega(r)=\lambda(r).

Application to periodic solutions of differential equations

Spectral theory can be seen as the infinite dimensional generalization of classical linear algebra. Linear differential operators and linear partial differential operators furnish prominent examples of such operators. The theorems of Hadamard–Lévy type have been applied to solve nonlinear differential equations.

I just give an example here, to give an idea of how this works, and also because I am quite lazy enough to check the details.

Following Brown & Lin (1979), we consider the Newtonian equation of motion:  u(t)+G(u(t))=p(t) u''(t) + \nabla G (u(t)) = p(t) where GG represents the ambiant potential, assumed to be smooth enough, and p ⁣:RRnp\colon \mathbf R\to\mathbf R^n is some external control. The problem studied by Brown and Lin is to prove the existence of periodic solutions when pp is itself periodic. The method consists in interpreting the left hand side as a non linear map defined on the Sobolev space EE of 2π2\pi-periodic C1\mathscr C^1-functions with a second derivative in F=L2([0;2π];Rn)F=L^2([0;2\pi];\mathbf R^n), with values in FF. Write LL for the linear operator uuu\mapsto u'' and NN for the (nonlinear) operator uG(u)u\mapsto \nabla G(u). Then LL is linear continuous (hence L(u)(v)=L(v)L'(u)(v)=L'(v)), and NN is continuously differentiable, with differential given by  N(u)(v)=(tQ(u(t))(v(t))) N'(u) (v) = \left( t \mapsto Q (u(t)) (v(t)) \right) for u,vEu,v\in E, and QQ is the Hessian of GG.

In other words, the differential (L+N)(u)(L+N)'(u) is the linear map vL(v)+Q(u(t))vv\mapsto L(v) + Q(u(t)) v. It is invertible if the eigenvalues of Q(u(t))Q(u(t)) are away from integers. Concretely, Brown and Lin assume that there are two constant symmetric matrices AA and BB such that AQ(x)BA\leq Q(x) \leq B for all xx, and whose eigenvalues λ1λn\lambda_1\leq \dots\lambda_n and μ1μn\mu_1\leq\dots\leq \mu_n are such that there are integers N1,,NnN_1,\dots,N_n with Nk2<λkμk<(Nk+1)2N_k^2<\lambda_k\leq\mu_k<(N_k+1)^2 for all kk. Using spectral theory in Hilbert spaces, these conditions imply that the linear operator L+Q(u) ⁣:EFL+Q(u)\colon E\to F is an isomorphism, and that (L+Q(u)1\|(L+Q(u)^{-1}\| is bounded from above by the constant expression  c=sup1knsup(λkNk2)1,((Nk+1)2μk)1). c= \sup_{1\leq k\leq n} \sup (\lambda_k-N_k^2)^{-1},((N_k+1)^2-\mu_k)^{-1} ).

Thanks to this differential estimate, the theorem of Hadamard–Lévy implies that the nonlinear differential operator L+NL+N is a global diffeomorphism from EE to FF. In particular, there is a unique 2π2\pi-periodic solution for every 2π2\pi-periodic control function pp.

I thank Thomas Richard for his comments.