\documentclass[reqno]{amsart}
\usepackage{hyperref}
\usepackage{graphicx,amssymb}


\AtBeginDocument{{\noindent\small
Variational and Topological Methods:
Theory, Applications, Numerical Simulations, and Open Problems (2012).
{\em Electronic Journal of Differential Equations},
Conference 21 (2014),  pp. 129--147.
ISSN: 1072-6691.  http://ejde.math.txstate.edu,
http://ejde.math.unt.edu \newline ftp ejde.math.txstate.edu}
\thanks{\copyright 2014 Texas State University - San Marcos.}
\vspace{9mm}}

\begin{document} \setcounter{page}{129}
\title[\hfilneg EJDE-2014/Conf/21 \hfil
Convergence of a mountain pass type algorithm]
{Convergence of a mountain pass type algorithm
  for strongly indefinite problems and systems}

\author[C. Grumiau, C. Troestler \hfil EJDE-2014/Conf/21\hfilneg]
{Christopher Grumiau, Christophe Troestler}  % in alphabetical order

\address{Christopher Grumiau \newline
  Institut Complexys, D{\'e}partement de Math{\'e}matique,
  Universit{\'e} de Mons, 20, Place du Parc B-7000 Mons Belgium}
\email{christopher.grumiau@umons.ac.be}

\address{Christophe Troestler \newline
  Institut Complexys, D{\'e}partement de Math{\'e}matique,
  Universit{\'e} de Mons, 20, Place du Parc B-7000 Mons Belgium}
\email{christophe.troestler@umons.ac.be}

\thanks{Published February 10, 2014.}
\subjclass[2000]{35J20, 58E05, 58E30, 35B38}
\keywords{Mountain Pass Algorithm; minimax; steepest descent method;
\hfill\break\indent   Schr\"odinger equation; spectral gap;
      strongly indefinite functional; ground state solutions;
\hfill\break\indent      Nehari manifold; systems}

\begin{abstract}
  For  a functional $\mathcal{E}$ and a peak selection that picks up a global
  maximum of $\mathcal{E}$ on varying cones, we study the convergence up to
  a subsequence to a critical point of the sequence generated by a
  mountain pass type algorithm. Moreover,
  by carefully choosing stepsizes, we establish the convergence of the whole
  sequence under a ``localization'' assumption on the critical point.
  We illustrate our results with two problems: an indefinite
  Schr\"odinger equation   and a superlinear Schr\"odinger system.
\end{abstract}

\maketitle
\numberwithin{equation}{section}
\newtheorem{theorem}{Theorem}[section]
\newtheorem{lemma}[theorem]{Lemma}
\newtheorem{proposition}[theorem]{Proposition}
\newtheorem{definition}[theorem]{Definition}
\newtheorem{remark}[theorem]{Remark}
\newtheorem{algorithm}[theorem]{Algorithm}
\allowdisplaybreaks

\section{Introduction}

Let us consider ${\mathcal H}$ a Hilbert space with inner product
$\langle{\cdot}|{\cdot}\rangle$ and norm $\|\cdot\|$, and a functional
$\mathcal{E}\in {\mathcal C}^1({\mathcal H}; {\mathbb R})$.
%
In this work, we develop a provably convergent ``general'' mountain pass type
algorithm to approximate saddle points of $\mathcal{E}$,
with a Morse index possibly larger than one.
%
The pioneer work in this direction is due to  Y.~S.~Choi and
P.~J.~McKenna~\cite{mckenna} who proposed a
constrained steepest descent method to compute saddle points with
one ``descent direction'' (such as a mountain pass solution).
A proof of convergence of a variant of that algorithm
was later given by Y.~Li and J.~Zhou in~\cite{zhou1,zhou2}.
To briefly describe it,
let us fix a closed subspace $E$ of ${\mathcal H}$ and $\varphi$
a continuous $E^\perp$-peak selection,
i.e.\ $\varphi(u)$ is the location of a maximum of
$\mathcal{E}$ on
$E \oplus {\mathbb R}^+ u := \{ e + ty \mid e \in E,\  t \geqslant 0\}$
for any $u\in {\mathcal H}\setminus E$ and $\varphi$ is constant on
$E \oplus {\mathbb R}^+ u$.
As it will be convenient in the rest of the paper that $\varphi$ is
not solely defined on a unit sphere, we present here a slightly
different version~\cite{tt}.

\begin{algorithm}[Mountain Pass Algorithm]   \label{mpa} \rm
 \quad
  \begin{enumerate}
  \item[(i)] Choose $u_0\in\operatorname{Ran} \varphi$, $\varepsilon >0$ and
 $n\leftarrow 0$;
  \item[(ii)] if $\|\nabla \mathcal{E}(u_n)\| \leqslant \varepsilon$
    then  stop; \\
    else compute
    \begin{equation*}
      u_{n+1}
      = \varphi\Big(u_n - s_n
        \frac{\nabla \mathcal{E}(u_n)}{\|\nabla \mathcal{E}(u_n)\|}\Big),
    \end{equation*}
    for some $s_n\in S(u_n) \subseteq ({0, +\infty})$
    where $S(u_n)$ is a set of ``admissible stepsizes''
    chosen so that at least the following inequality
    (which is reminiscent of \emph{Armijo condition}~\cite{Nocedal-Wright06})
    holds:
    \begin{equation*}
      \mathcal{E}(u_{n+1})-\mathcal{E}(u_n)
      < - \tfrac{1}{2} s_n \|{\nabla \mathcal{E}(u_n)}\|;
    \end{equation*}
  \item[(iii)] let $n\leftarrow n+1$ and go to step $2$.
  \end{enumerate}
\end{algorithm}

Y.~Li and J.~Zhou proved that $(u_n)$ converges
to a nontrivial critical point of $\mathcal{E}$ up
to a subsequence.
The proof of convergence is performed in the space ${\mathcal H}$
to ensure that the rate of
convergence for the discretized problem does not deteriorate
when the approximating subspace becomes finer.
%
The original goal of the authors for
introducing $E$  was to
try to obtain multiple critical points by taking $E$
as the linear subspace generated by previously found solutions
which the algorithm must
try to avoid.  The proof is performed in two steps.
First, they show
that $s_n$  exists and that $\mathcal{E}$ decreases along $(u_n)_{n\in\mathbb{N}}$.
This step relies on the following Deformation Lemma.

\begin{lemma}  \label{Udeflemma}
  If $\varphi$ is continuous, $u_0\in \operatorname{Ran} \varphi$,
$u_0\notin E$, $\nabla  \mathcal{E}(u_0)\neq 0$,
  then there exists $s_0>0$ such that
  \begin{equation*}
    \forall s \in ({0, s_0}],\quad
    \mathcal{E}\bigl(\varphi(u_{s})\bigr) - \mathcal{E}(u_0)
    < - \tfrac{1}{2} s \mathopen\|\nabla \mathcal{E}(u_0)\|,
  \end{equation*}
  where
  \begin{equation*}
    u_{s}
    := u_0 - s \frac{\nabla \mathcal{E}(u_0)}{\|\nabla \mathcal{E}(u_0)\|} .
  \end{equation*}
\end{lemma}

The second step consists in proving,
under some traditional assumptions on $\varphi$, that
a subsequence of $(u_n)$ converges.  For this, it is essential to show
that the stepsize $s_n$ controls the distance between $u_n$ and $u_{n+1}$
and that $s_n$ is chosen in such a way that it is close to $0$
only when ``mandated'' by the functional.  Let us remark that
the choice of $\varphi$ is very
sensitive. Indeed,  to seek  sign-changing
critical points, the modified
Mountain Pass algorithm was introduced by
J.~M.~Neuberger~\cite{Neuberger97} (see also~\cite{cdn}). He
considers  Algorithm~\ref{mpa} above and only modifies the projection $\varphi$
into a ``sign-changing peak selection'' $\varphi_N$ which is a
map defined from
the set of sign-changing functions of
${\mathcal H}\setminus\{0\}$ to ${\mathcal H}\setminus\{0\}$ such that, for any
$u$, $\mathcal{E}\bigl(\varphi_N(u)\bigr)>0$ and $\varphi_N(u)$ is a maximum of $\mathcal{E}$ on
${\mathbb R}^+ u^+ \oplus {\mathbb R}^+ u^-$ where $u^+(x):= \max\{0,u(x)\}$ and
$u^-(x):=\min\{0, u(x)\}$.  Although in practice it appears to converge to a
nontrivial sign-changing critical point, its
convergence has yet to be formally proved.

In this paper, for any $u$, $\varphi(u)$ is allowed to
be any maximum point of $\mathcal{E}$ inside an
abstract cone $C_u$
and we are interested in giving
assumptions on $C_u$ which imply the convergence of the
mountain pass algorithm. This work is partly motivated by the
article~\cite{noris} where
the authors define the notion of ``natural
constraints'' to seek nontrivial
critical points of functionals.
%
To the best of the authors' knowledge, using peak selections along
cones that vary from point to point in a ``continuous'' fashion
(see~\eqref{Ac2}) is novel and allows us to propose a general,
provably convergent, algorithm that subsumes into a unified setting
several previous mountain-pass type algorithms.  We will show how it
can be applied both to strongly indefinite problems and to systems.

Let us first make precise the peak selection $\varphi$ that we use.
We write
$\operatorname{int}C$ for the interior of $C$ relative to
$\operatorname{\overline{span}} C$,
the smallest closed subspace containing $C$,
for the topology induced by~${\mathcal H}$.

\begin{definition}  \label{defpeak} \rm
  Let ${\mathcal A}$ be an open subset of~${\mathcal H}$.
  We say that $\varphi$ is a \emph{peak selection for $(C_u)_{u \in {\mathcal A}}$}
  if $\varphi$ is a map from ${\mathcal A}$ to ${\mathcal A}$ such
  that, for all $u \in {\mathcal A}$,
  \begin{itemize}
  \item[(i)] $C_u$ is a closed cone pointed at~$0$;
    \vspace{0pt plus 1pt}
  \item[(ii)] $\varphi(u) \in \operatorname{int} C_u$;
    \vspace{0pt plus 1pt}
  \item[(iii)] \label{phi-uniqueness} for any $v\in \operatorname{int}C_u$,
    $\varphi(v)= \varphi(u)$;
    \vspace{0pt plus 1pt}
  \item[(iv)] $\varphi(u)$ is a global maximum point of $\mathcal{E}$ on $C_u$.
  \end{itemize}
\end{definition}

Note that properties (ii) and (iii) imply that $\varphi(\varphi(u)) =
\varphi(u)$.
We write that $d\perp C_u$ if and only
if $d\perp \operatorname{\overline{span}} C_u $
for the inner product $\langle{\cdot}|{\cdot}\rangle$.
In Section \ref{conmpa}, we assume that
$\varphi$ is continuous and
that  $(C_u)$ verifies the following conditions:
\begin{equation}
\forall u\in {\mathcal A}, \ u\in C_u
  %\tag{$AC_1$}
\label{Ac1}\\
\end{equation}
and
\begin{equation}  %\tag{$AC_2$} 
\label{Ac2}
  \begin{gathered}
    \exists \gamma \in \bigl({0,\frac{\pi}{2}}\bigr), \;
    \exists \delta \in (0,1),\  \forall u_0\in \operatorname{Ran}\varphi,\;
    \exists r>0,\;
    \forall \tilde{u}_0\in \operatorname{Ran}\varphi \cap B(u_0,r),  \\
       \forall d \in B(0,r),\ d \perp C_{\tilde{u}_0},\quad
    C_{\tilde{u}_0 + d} \cap B(u_0, r)\subseteq C_{\tilde{u}_0}
    + [1-\delta, 1+\delta] \,  A_\gamma(d),
  \end{gathered}
\end{equation}
where $A_\gamma(d) := \{ d' \mid \|d'\| = \|d\|$ and
$\angle (d',d)\leqslant \gamma\}$ and
$\angle (d,d') := \arccos \bigl(\frac{\langle{d}|{d'}\rangle}{\|d\mathclose\|
\mathopen \|d'\mathclose\|}\bigr)$
denotes the angle between two non-zero vectors $d$ and $d'$
(we set $A_\gamma(0) := \{0\}$).
This assumption, which speaks about the behavior of the cones under
small deformations, is essential to prove a Deformation Lemma in this
generalized setting
(see Lemma~\ref{deflemma}).
This lemma ensures the non-emptiness of the
set $S(u_0)$ of admissible stepsizes at $u_0$ which we now define.
For any $u_0\in \operatorname{Ran} \varphi$ such that
$\nabla \mathcal{E}(u_0)\ne 0$, we set
\begin{align*}
& S^{*}(u_0) \\
&:= \Bigl\{ s>0 :  u_s := u_0 -
  s\frac{\nabla\mathcal{E}(u_0)}{\|\nabla\mathcal{E}(u_0)\|} \in {\mathcal A}
  \text{ and }
  \mathcal{E}\bigl(\varphi(u_s)\bigr) - \mathcal{E}(u_0)
  < -\alpha s \mathopen\|\nabla \mathcal{E}(u_0)\|  \Bigr\}
\end{align*}
for some value $\alpha >0$ given  by  Lemma~\ref{deflemma}  and  we require
that $s\in S(u_0):= S^{*}(u_0)\cap \bigl[{\frac{1}{2}\sup  S^{*}(u_0), +\infty}\bigr)$.  Other definitions of
admissible stepsizes are possible provided they imply a local
uniformity in the sense that $s_n \in S(u_n)$ forces the stepsize $s_n$ not
to be small when the gradient is not (see Lemma~\ref{convlemma1}).
The definition given above
draws its inspiration from a
paper~\cite{tt} written by N.~Tacheny and C.~Troestler.

To obtain the convergence of $(u_n)$ up to a subsequence
(see Theorem~\ref{convthm}), we unfortunately need to
replace~\eqref{Ac2} with the following stronger assumption:
\begin{equation} %  \tag{$AC_3$}
\label{Ac-xi}
  \left.\begin{minipage}{0.83\linewidth}
      there exist a closed subspace $E\subseteq {\mathcal H}$ (possibly
      infinite dimensional) and a family of
      ${\mathcal C}^1$-vector fields $\xi_i :{\mathcal A} \to E^{\perp}$,
      $i=1,\dots,k$, for some $k\in\mathbb{N}$, such that
      for all $u \in {\mathcal A}$ and $i \in \{1,\dotsc,k\}$,
      \begin{enumerate}
      \item[(i)] the family $\bigl(\xi_i(u)\bigr)_{i=1}^k$ is orthonormal;
        \vspace{0.5ex}
      \item[(ii)] $\forall v \in V_u,\ \xi'_i(u)[v] \in V_u$,
        where $V_u := \operatorname{\overline{span}}
        \{\xi_1(u), \dotsc, \xi_k(u)\}$;
        \vspace{0.5ex}
      \item[(iii)] $\forall v\in \operatorname{int} C_u\cap{\mathcal A}$,
        $\xi_i(v) = \xi_i(u)$;
        \vspace{0.5ex}
      \item[(iv)] $\langle{u}|{\xi_i(u)}\rangle \ne 0$;
        \vspace{0.5ex}
      \item[(v)] $\exists r > 0$, $\xi'_i$ is bounded on
        $\{ u \mid \operatorname{dist}(u, \operatorname{Ran} \varphi) < r \} \cap {\mathcal A}$.
      \end{enumerate}
      For $u \in {\mathcal A}$, the cone $C_u$ is defined as
      $C_u := E \oplus \{ \sum_i t_i \xi_i(u) \mid
      t_i \geqslant 0 \text{ for all } i \}$.
    \end{minipage} \ \right\}
\end{equation}
Here, the notation $\operatorname{dist}(u,\partial{\mathcal A})$ stands for
$\inf\{ \|u - v\| \mid v \in \partial{\mathcal A} \}$.
%
Let us remark that conditions~\eqref{Ac-xi} (i), (ii), (iv) and (v)
are already present (albeit somehow implicitly for~(iv))
in~\cite{noris} in the context of trivial ${\mathcal C}^1$-subbundles intead of
cones.
The additional condition~(iii) is equivalent to
$\forall v \in \operatorname{int} C_u\cap{\mathcal A},\ C_v = C_u$.
%
This condition is rather natural to require in view of
property~(iii) of the definition of peak selection.
This ``finite presentation'' of the
cones is used  in Lemma~\ref{convlemma2-xi} to ensure that the stepsize $s_n$
controls the distance between $u_{n+1}$ and~$u_n$.

As a particular case of~\eqref{Ac-xi}, let us mention that we can
work with a family of continuous linear projectors
(see Proposition~\ref{conv-P_i}).  This
case is an abstract formulation of the setting of~\cite{sym2,sym1}
where the convergence (up
to a subsequence) of a mountain pass type algorithm for systems
has been announced.

In Section~\ref{convmpa2}, we are interested in the convergence of the whole
sequence generated by the Mountain Pass Algorithm.  To that aim, we
need to refine the definition of $S^*$  in order to control
$\mathcal{E}\bigl(\varphi(u_n - s \frac{\nabla u_n} {\|\nabla u_n\|})\bigr)$
for any $0<s<s_n$.

In Section~\ref{applmpa}, we illustrate our method with  two semi-linear
problems. The first application takes its inspiration from a paper due to
A.~Szulkin and T.~Weth~\cite{szulkin} in which the authors study the following
Schr\"odinger problem
\begin{equation}
  \label{eq:Szulkin}
    \begin{gathered}
      -\Delta u(x) + V(x) u(x)= |u(x)|^{p-2}u(x),\quad x\in\Omega, \\
      u(x)=0,\quad  x \in\partial\Omega ,
    \end{gathered}
\end{equation}
where $V:\Omega\to {\mathbb R}$ is such that $0$ is in a spectral gap
of $-\Delta + V$ and $2<p<2^*:=\frac{2N}{N-2}$
($+\infty$ when $N=2$).  They are interested in the existence of
non-zero solutions on an open bounded domain $\Omega\subseteq {\mathbb R}^N$ or
on $\Omega = {\mathbb R}^N$ (in the latter case, $V$ is assumed to be
 $1$-periodic in each $x_i$, $i=1,\dotsc,N$).  Solutions to this equation are
critical points of the indefinite functional
\begin{equation}
  \label{functional-indefinite}
  \mathcal{E}: {\mathcal H}\to {\mathbb R} : u\mapsto \frac{1}{2}\int_\Omega\left(
    |\nabla u(x)|^2 + V(x) u(x)^2\right) \,{\mathrm d} x
  -\frac{1}{p}\int_\Omega |u (x)|^p\,{\mathrm d} x,
\end{equation}
where ${\mathcal H}= H^1_0 (\Omega)$. The first proof of the existence of
non-zero critical points for $\mathcal{E}$ when $-\Delta + V$ is not positive
definite and $\Omega$ is an open bounded domain is due to
P.~H.\ Rabinowitz~\cite{rabin}.  Recently,
A.~Szulkin and T.~Weth proposed an alternative method~\cite{szulkin}
that also makes easier to deal with
the lack of compactness that occurs when $\Omega = {\mathbb R}^N$.
 Denoting $E$ the negative eigenspace of $-\Delta + V$, they introduce
the following nonlinear map
\begin{equation*}
  \varphi: {\mathcal H}\setminus E\to {\mathcal H}: u\mapsto \varphi(u)
\end{equation*}
where $\varphi(u)$ is the point at which
$\mathcal{E}$ reaches its maximum value on $E \oplus {\mathbb R}^+ u$.
They prove that minimizing $\mathcal{E}$ on
 $\operatorname{Ran} \varphi = \bigl\{ u\in {\mathcal H}\setminus
E \bigm| \partial\mathcal{E}(u)[v]=0$
for $v=u$ and any $v\in E \bigr\}$ yields a non-zero
solution with least energy.
%
Notice that, here, $E$  is used to deal with
the indefiniteness of the problem
and not to compute multiple critical points as in
the papers of J.~Zhou \& al.~\cite{zhou1,zhou2}.  We will
prove that our algorithm converges for this problem.  The numerical
solutions that we obtain lead to some conjectures on
the symmetries of ground state solutions.


The second application is based on a paper by
B.~Noris and G.~Verzini \cite{noris}. The
authors study the superlinear Schr\"odinger system
\begin{equation}
  \label{eqNoris-general}
    \begin{gathered}
      -\Delta u_i(x) = \partial_iF\bigl(u_1(x),\dots,u_k(x)\bigr),
      \quad x\in\Omega, \\
      u_i(x)=0,\quad x \in\partial\Omega ,
    \end{gathered}
 \quad i=1,\dots,k,
\end{equation}
where $k \in \mathbb{N}$. They require that  $\Omega\subseteq {\mathbb R}^N$ is
a bounded smooth
domain and $F\in {\mathcal C}^2({\mathbb R}^k; {\mathbb R})$.
Note that the system $-\Delta u_i =\mu_i
u_i^3 + u_i \sum_{j\neq i} \beta_{i,j} u_j^2$ where $\mu_i>0$
and $\beta_{i,j} = \beta_{j,i}$ is a particular case
of~\eqref{eqNoris-general}.  Such type of nonlinearities have been studied
due to their applications to nonlinear optics and to Bose-Einstein
condensation (see~\cite{conti,terracini2,dancer,terracini}).
Solutions to~\eqref{eqNoris-general} are
critical points of the functional
\begin{equation}
  \label{functional-system}
  \mathcal{E}: {\mathcal H}\to {\mathbb R} : u = (u_1,\dotsc, u_k) \mapsto
  \frac{1}{2}\int_\Omega |\nabla u(x)|^2
  \,{\mathrm d} x - \int_\Omega F(u)\,{\mathrm d} x,
\end{equation}
where ${\mathcal H} = H^1_0 (\Omega; {\mathbb R}^k)$.
As already mentioned, B.~Noris and G.~Verzini~\cite{noris}
propose a general method of ``natural constraints''.
Applied to the above problem, it goes as follows.
Denote ${\mathcal A} := \{u\in {\mathcal H} :u_i\neq 0 \text{ for every } i\}$.
To find a solution
$u=(u_1,\dots, u_k)$ of~\eqref{eqNoris-general}
with $u_i \ne 0$ for all $i=1,\dots,k$,
they minimize $\mathcal{E}$ on the constraint
${\mathcal N} := \bigl\{u\in {\mathcal A}
\bigm| \forall  i=1,\dots,k,\ \int_{\Omega}
|\nabla u_i|^2\,{\mathrm d} x = \int_{\Omega} \partial_i F(u)u_i\,{\mathrm d} x
\bigr\}$.   We will show that, under their assumptions,
${\mathcal N} =\operatorname{Ran} \varphi$ with $\varphi$ being the peak
selection
\begin{equation*}
  \varphi : {\mathcal A} \to {\mathcal A} : u \mapsto
  \operatorname{argmax} \bigl\{ \mathcal{E}(t_1 u_1,\dots, t_k u_k) :
  t_1>0, \dotsc, t_k > 0 \bigr\} .
\end{equation*}
Again, we prove that our algorithm converges for this problem and
some numerical experiments are performed.

\section[Steepest descent method]{Steepest
  descent method on varying cones}
\label{conmpa}

\subsection{Uniform Deformation Lemma}

Let $\mathcal{E} : {\mathcal H} \to {\mathbb R}$ be a ${\mathcal C}^1$-functional
defined on a Hilbert space
${\mathcal H}$ and ${\mathcal A}$ an open subset of~${\mathcal H}$.
The following lemma is
instrumental in proving the convergence of the algorithm.

\begin{lemma}[Uniform Deformation Lemma]
  \label{deflemma}
  Let $\varphi : {\mathcal A} \to {\mathcal A}$ be a peak selection
for $(C_u)_{u \in {\mathcal A}}$ and
  $u_0\in \operatorname{Ran}\varphi$ be such that $\nabla \mathcal{E}(u_0) \ne 0$.
  Assume that $\varphi$ is continuous at~$u_0$
  and that \eqref{Ac1}--\eqref{Ac2} hold.
  Then there exist $s_0>0$ and $r_0 >0$ such that,
  for any $s \in ({0, s_0}]$ and
  $\tilde{u}_0\in B(u_0, r_0)\cap \operatorname{Ran} \varphi$, one has
  \begin{itemize}
  \item $\nabla \mathcal{E}(\tilde{u}_0)\neq 0$,
  \item $\tilde{u}_s\in {\mathcal A}$ where $\tilde{u}_s:= \tilde{u}_0
    -s\frac{\nabla \mathcal{E}(\tilde{u}_0)}{
      \|\nabla \mathcal{E}(\tilde{u}_0)\|}$ and
  \item there exists some $\alpha > 0$ solely depending on $\gamma$
    and $\delta$ given in assumption~\eqref{Ac2} such that
    \begin{equation}
      \label{step}
      \mathcal{E}\bigl(\varphi(\tilde{u}_s)\bigr) - \mathcal{E}(\tilde{u}_0)
      < -\alpha s \mathopen\|\nabla \mathcal{E}(\tilde{u}_0)\|.
    \end{equation}
  \end{itemize}
\end{lemma}

\begin{proof}
  Let $u_0 \in \operatorname{Ran}\varphi \subseteq {\mathcal A}$ and
  let us consider $\gamma$, $\delta$
  and $r$ given by the assumption~\eqref{Ac2} for $u_0$.
  Since ${\mathcal A}$ is open, there exists
  $\varepsilon_1 >0$ such that for any $u\in B(u_0, \varepsilon_1)$ and
  $v\in B(u,\varepsilon_1)$,
  one has $u,v\in {\mathcal A}$, $\nabla\mathcal{E}(u)\ne 0$,
  $\nabla\mathcal{E}(v)\ne 0$ and $u,v\in B(u_0, r)$.

  For any $u\in B(u_0, \varepsilon_1)$, let $d_u := -{\nabla \mathcal{E}
  (u)}/{\|\nabla \mathcal{E}(u)\|}$.  Then
  \begin{equation*}
    \forall u\in B(u_0, \varepsilon_1),\
    \forall d\in A_\gamma(d_u),\quad
    \langle{\nabla\mathcal{E}(u)}|{d}\rangle \leqslant -\cos\gamma\,
    \|\nabla\mathcal{E}(u)\|.
  \end{equation*}
  Let $\tilde\gamma := \frac{1}{2}\cos\gamma > 0$.
  Taking $\varepsilon_1$ smaller if necessary, we may assume that
  \begin{equation*}
    \forall u, v \in B(u_0, \varepsilon_1),\
    \forall d\in A_\gamma(d_u),\quad
    \langle{\nabla\mathcal{E}(v)}|{d}\rangle
    < -\tilde\gamma \mathopen\|\nabla\mathcal{E}(u)\|.
  \end{equation*}
  Thus, on one hand, there exists $\varepsilon_2>0$ such that,
  for any $u\in
  B(u_0, \varepsilon_2)$, $v \in B(u,\varepsilon_2)$, $d\in
  A_\gamma(d_u)$ and $\sigma \in ({0, \varepsilon_2})$,
  \begin{equation*}
    \langle{\nabla\mathcal{E}(v + \sigma d)}|{d}\rangle
    < - \tilde\gamma \mathopen\|\nabla\mathcal{E}(u)\|.
  \end{equation*}
  For any  $\tilde{u}_0\in B(u_0, \varepsilon_2)\cap
  \operatorname{Ran} \varphi$, $v\in C_{\tilde{u}_0}\cap B(\tilde{u}_0,
  \varepsilon_2)$,  $d\in A_\gamma(d_{\tilde{u}_0})$ and
  $\sigma < \varepsilon_2$, the Mean Value Theorem implies there
  exists a $\tilde\sigma \in ({0,\sigma})$ such that
  \begin{align}
    \mathcal{E}(v  + \sigma d) - \mathcal{E}(\tilde{u}_0)
    &\leqslant \mathcal{E}(v+ \sigma d) -\mathcal{E}(v)  \label{eq:use max E}\\
    &= \bigl\langle{\nabla \mathcal{E}( v+ \tilde\sigma d)}
    \,{\bigm|}\,{\sigma d}\bigr\rangle \notag\\
    &< -\tilde\gamma \sigma \mathopen\|\nabla \mathcal{E}(\tilde{u}_0)\|,
    \label{eqlem}
  \end{align}
  where the first inequality results from $v\in C_{\tilde{u}_0}$
  and $\tilde{u}_0 = \varphi(\tilde u_0)$ is a global maximum of $\mathcal{E}$
  on~$C_{\tilde{u}_0}$.

  On the other hand, by the continuity of $\varphi$ at~$u_0$, there exist
  $s_0 \in ({0,r})$
  and $\varepsilon_3 \in \bigl({0,
    \min\{r, \tfrac{1}{3}\varepsilon_2\}}\bigr)$ such that,
  for any $\tilde{u}_0\in B(u_0,\varepsilon_3)$
  and $s \in [{0,s_0}]$, one has
  $\varphi(\tilde{u}_0 + s d_{\tilde{u}_0})
  \in B(u_0,\min\{r, \tfrac{1}{3}\varepsilon_2\})$.
  Let $\tilde{u}_s := \tilde{u}_0 + s d_{\tilde{u}_0}$.
  If in addition $\tilde u_0 \in \operatorname{Ran}\varphi$, one has
  $d_{\tilde{u}_0} \perp \operatorname{\overline{span}}
  C_{\tilde{u}_0}$ (because $\tilde{u}_0 =
  \varphi(\tilde{u}_0) \in \operatorname{int} C_{\tilde{u}_0}$ is a local
  maximum)
  and therefore one deduces from assumption~\eqref{Ac2} that
  \begin{equation*}
    \varphi(\tilde{u}_s)\in C_{\tilde{u}_s} \cap B(u_0,r)
    \subseteq C_{\tilde{u}_0}
    + [{1-\delta,1+\delta}] A_{\gamma}(sd_{\tilde{u}_0}).
  \end{equation*}
  Thus, $\varphi(\tilde{u}_s) = v_s + K_s s d^*_s$
  for some $v_s\in C_{\tilde{u}_0}$,
  $K_s\in [{1-\delta, 1+\delta}]$ and $d^*_s\in
  A_{\gamma}(d_{\tilde{u}_0})$. So, possibly taking $s_0$
  smaller, we get that $K_s s < \tfrac{1}{3}\varepsilon_2$ and
  $v_s = \varphi(\tilde{u}_s) -
  K_s s d^*_s \in B(\tilde{u}_0,\varepsilon_2)$.
  Using Equation~\eqref{eqlem}, we conclude that
  \begin{equation*}
    \mathcal{E}\bigl(\varphi(\tilde{u}_s)\bigr) - \mathcal{E}(\tilde{u}_0)
    = \mathcal{E}(v_s + K_s s d^*_s) - \mathcal{E}(\tilde{u}_0)
    \leqslant -\tilde\gamma\, (1 -\delta) s
    \mathopen\|\nabla\mathcal{E}(\tilde{u}_0)\|
  \end{equation*}
  for any $\tilde{u}_0\in B(u_0,\varepsilon_3)\cap \operatorname{Ran} \varphi$
  and $s \in ({0, s_0}]$.
\end{proof}

\begin{remark}\label{uniform1} \rm \quad
 \begin{itemize}
\item Equation~\eqref{eq:use max E} is the unique place we use  that
  $\varphi(u)$ is a global maximum of $\mathcal{E}$ on $C_u$.  This assumption can be
  weakened by only requiring that the neighborhood on which $\varphi(u)$
  achieves the maximum of $\mathcal{E}$ is locally uniform w.r.t.\ $u$:
  \begin{equation}
    \label{eq:local-max-phi}
    \forall u_0 \in \operatorname{Ran}\varphi,\
    \exists \rho >0,\
    \forall u \in \operatorname{Ran} \varphi \cap B(u_0, \rho),\quad
    \mathcal{E}\bigl(\varphi(u)\bigr) = \max_{v \in C_u \cap B(u, \rho)} \mathcal{E}(v).
  \end{equation}
  This assumption allows the existence of multiple maximum points
  in~$C_u$.  It was not used in Definition~\ref{defpeak} for
  simplicity but also because, in the examples of
  Section~\ref{applmpa}, $\varphi(u)$ is a maximum on the whole~$C_u$.

\item Let us also note that, if we are just interested in the
  inequality~\eqref{step} at $u_0$ (and not for all $\tilde u_0$ in a
  neighborhood of $u_0$), we only need to require that
  $\varphi(u)$ is a local maximum of $\mathcal{E}$ on~$C_u$.

\item A careful reader may notice that we did not really use the fact
  that $C_u$ is a cone pointed at~$0$.  However, if $(C_u)$ was just a
  family of sets satisfying
  \eqref{Ac1}, \eqref{Ac2}, \eqref{eq:local-max-phi} and the fact that
  $\varphi(u) \in \operatorname{int} C_u$ in a locally uniform way:
  \begin{equation}
    \label{eqrem}
    \forall u_0 \in \operatorname{Ran}\varphi,\ \exists \rho > 0,\
    \forall u \in \operatorname{Ran} \varphi \cap B(u_0, \rho),\quad
    B(u, \rho) \cap \operatorname{\overline{span}} C_u \subseteq C_u \text{,}
  \end{equation}
  then the cone $\hat C_u$,
  defined as the closure of $\{ tv \mid t \geqslant 0$ and $v \in
  C_u\}$, also satisfies \eqref{Ac1}, \eqref{Ac2}, \eqref{eq:local-max-phi}
  and $\varphi(u) \in \operatorname{int} \hat C_u$.  So very little is
  gained by not using cones, especially because they are the natural
  structures encountered in our examples.

\item As a consequence of the above Deformation Lemma, one can interpret
  $\operatorname{Ran}\varphi$ as
  somewhat a natural constraint for $\mathcal{E}$ in the sense of~\cite{noris}.
  More precisely, it implies that
  if $u_0\in \operatorname{Ran} \varphi$ is a local minimum of $\mathcal{E}$ on $\operatorname{Ran}
  \varphi$ then $u_0$ is a critical point of $\mathcal{E}$ on the whole space~${\mathcal H}$.
\end{itemize}
\end{remark}


\subsection{Convergence up to a subsequence}

In this section, we first remark that
it is possible to
construct a sequence of stepsizes $s_n$ such that the energy $\mathcal{E}$ decreases
along the sequence $(u_n)_{n\in\mathbb{N}}$ generated by
Algorithm~\ref{mpa}. In the following,  without loss of
generality, we can assume that $\nabla \mathcal{E}(u_n)\neq 0$ for  any
$n\in\mathbb{N}$ (otherwise the algorithm finds a critical point
in a finite number of steps).


\begin{proposition}  \label{stability}
  If   $s_n >0$ verifies inequality~\eqref{step} given in
  Lem\-ma~\ref{deflemma} for any $n\in\mathbb{N}$,
  then the functional $\mathcal{E}$ decreases along the sequence
$(u_n)_{n\in\mathbb{N}}$.
\end{proposition}

\begin{proof}
  As $\nabla \mathcal{E}(u_n)\neq 0$, $s_n$ is well-defined by
Lemma~\ref{deflemma}. By
  construction, we have
  \begin{equation*}
    \mathcal{E}(u_{n+1})-\mathcal{E}(u_n)
    = \mathcal{E}\Big(\varphi\bigl(u_n-s_n\frac{\nabla \mathcal{E}(u_n)}{
        \|\nabla \mathcal{E}(u_n)\|}\bigr)\Big)-\mathcal{E}(u_n)
    < -\alpha s_n  \mathopen\|\nabla \mathcal{E}(u_n)\|
    <0.
  \end{equation*}
  So, $\mathcal{E}(u_{n+1})<\mathcal{E}(u_n)$.
\end{proof}

As explained in the introduction, we now consider
the sets $S^*(u_0)$ and $S(u_0)$. The set $S^*(u_0)$ is
not empty as soon as $u_0$ is
not a critical point of $\mathcal{E}$ (thanks to the Deformation
Lemma). Concerning the set $S(u_0)$, it is not-empty once $\mathcal{E}$
is bounded from below on $\operatorname{Ran} \varphi$, an assumption that
we will later make (see Theorem~\ref{convtheorem}).

\begin{lemma}   \label{convlemma1}
  If
  $u_0\in \operatorname{Ran} \varphi$, $\nabla \mathcal{E}(u_0)\neq 0$ and  $\varphi$ is continuous at
  $u_0$, then there exists an open
  neighborhood $V$ of $u_0$ and $s^*>0$ such that $S(u)\subseteq [s^*,+\infty)$
  for any $u\in V\cap \operatorname{Ran} \varphi$.
\end{lemma}

\begin{proof}
  By the Uniform Deformation Lemma~\ref{deflemma}, there exists $s_0>0$ and
  $r_0>0$ such that, for any $0<s\leqslant s_0$ and $u\in B(u_0, r_0)\cap
  \operatorname{Ran}\varphi$, we have $u_s := u-s
  \frac{\nabla \mathcal{E}(u)}{\|\nabla\mathcal{E}(u)\|} \in {\mathcal A}$,
  $\nabla\mathcal{E}(u)\neq 0$ and
  \begin{equation}
    \label{step2}
    \mathcal{E}\bigl( \varphi(u_s) \bigr)-\mathcal{E}(u)
    < -\alpha s \mathopen\|\nabla \mathcal{E}(u)\|.
  \end{equation}
  In particular, for any $u\in B(u_0, r_0)\cap \operatorname{Ran}\varphi$, $s_0\in
  S^*(u)$. It follows that $S(u)\subseteq [{\frac{s_0}{2},+\infty})$. It
  suffices to take $s^* \leqslant s_0/2$.
\end{proof}

\begin{remark}   \label{uniform2} \rm
  To prove Lemma~\ref{convlemma1},  let us remark
  that we could only use
  inequality~\eqref{step2} at $u=u_0$ for $s = s_0$ fixed.
  Indeed, by continuity,  it directly
  implies  that $s_0 \in S(u)$ for $u$ close to $u_0$.  However, to
  obtain Lemma~\ref{convlemma1} for $\tilde S(u)$ (see
  Section~\ref{convmpa2}) instead of $S(u)$, the full strength of the
  Deformation Lemma is needed.
\end{remark}

From now on, we have to require condition~\eqref{Ac-xi}.
Let us first show it subsumes~\eqref{Ac2}.

\begin{lemma}   \label{orthog-derivative} \rm
  Let $(\xi_i)_{i=1}^k$ be the family of vector fields given by
  \eqref{Ac-xi} and assume \eqref{Ac1} holds.  Then
  \begin{equation*}
    \forall u \in {\mathcal A},\
    \forall d \perp C_u,\quad
    \sum_{i=1}^k  \bigl\langle{u}\,{\bigm|}\,{\xi_i(u)}\bigr\rangle
    \, \xi'_i(u)[d]
    = d - \sum_{i=1}^k \bigl\langle{u}\,{\bigm|}\,{\xi_i'(u)[d]}\bigr\rangle
    \, \xi_i(u).
  \end{equation*}
\end{lemma}

\begin{proof}
  For any $u \in {\mathcal A}$, the fact that $u \in C_u \subseteq V_u$ and that
  $(\xi_i)_{i=1}^k$ is an orthonormal basis of $V_u$ imply $u = \sum
  \langle{u}|{\xi_i(u)}\rangle \xi_i(u)$.  Differentiating in a direction
$d  \in {\mathcal H}$, yields
  \begin{equation*}
    d = \sum_{i=1}^k \langle{d}|{\xi_i(u)}\rangle \; \xi_i(u)
    + \sum_{i=1}^k \langle{u}|{\xi_i'(u)[d]}\rangle \, \xi_i(u)
    + \sum_{i=1}^k \langle{u}|{\xi_i(u)}\rangle \, \xi'_i(u)[d].
  \end{equation*}
  If $d \perp \operatorname{\overline{span}} C_u$, the first term
  vanishes.  This completes the proof.
\end{proof}

\begin{proposition}
  Properties~\eqref{Ac1} and~\eqref{Ac-xi} imply~\eqref{Ac2}.
\end{proposition}

\begin{proof}
  Let $\delta \in ({0,1})$ (property ~\eqref{Ac2} will be
  satisfied whatever value is chosen).  Simple geometrical
  considerations show that there exists a $\gamma \in ({0,
    \pi/2})$ such that
  \begin{equation*}
    B(d, \delta \|d\|) \subseteq
    [{1-\delta, 1+\delta}] A_\gamma(d).
  \end{equation*}
  Let $u_0\in \operatorname{Ran}\varphi$.  As
$u_0 \in \operatorname{int}  C_{u_0}$, there exist $\alpha > 0$ such that
  $\langle{u_0}|{\xi_i(u_0)}\rangle > \alpha$ for all $i$.
  Using the continuity of $\xi_i$ and $\xi'_i$ at $u_0$, we can choose
  $r$ sufficiently small and a $M > 0$ (depending only on $u_0$) so
  that, for all $u \in B(u_0,r)$ and all $i= 1,\dotsc, k$,
  \begin{align*}
    \langle{u}|{\xi_i(u)}\rangle > \alpha, \quad
    \|\xi_i(u) - \xi_i(\tilde{u}_0)\| \leqslant \varepsilon,\quad
    \|\xi'_i(u)\| \leqslant M,  \quad\text{and}\quad
    \|\xi'_i(u) - \xi'_i(\tilde{u}_0)\| \leqslant \varepsilon,
  \end{align*}
  where $\varepsilon >0$ is a constant depending only on $\delta$ and
  $u_0$ (to be chosen later).

  Let $\tilde u_0 \in B(u_0, r)$ and $d \in B(0,r)$ such that $d \perp
  C_{\tilde{u}_0}$.   Let $w \in
  C_{\tilde{u}_0 + d} \cap B(u_0, r)$.  One can write $w = e + \sum
  t_i \xi_i(\tilde{u}_0 + d)$ for some $e \in E$ and $t_i \geqslant
  0$. Let us start by noticing that $t_i =
  \langle{w}|{\xi_i(\tilde{u}_0 + d)}\rangle$.  Therefore, recalling that
  $\|\xi_i\| = 1$, one deduces
  \begin{align}
    \bigl|t_i - \langle{\tilde{u}_0}|{\xi_i(\tilde{u}_0)}\rangle\bigr|
    &\leqslant \bigl|\bigl\langle{w - \tilde{u}_0}\,{\bigm|}\,{
      \xi_i(\tilde{u}_0 + d)}\bigr\rangle \bigr|
    + \bigl| \bigl\langle{\tilde{u}_0}\,{\bigm|}\,{
        \xi_i(\tilde{u}_0 + d) - \xi_i(\tilde{u}_0)}\bigr\rangle \bigr|
      \notag\\
    &\leqslant \|w - \tilde{u}_0\| + \|\tilde{u}_0\|
    \|\xi_i(\tilde{u}_0 + d) - \xi_i(\tilde{u}_0)\|  \notag\\
    &\leqslant 2r + (\|u_0\| + r) \varepsilon .
    \label{eq:t-i}
  \end{align}
  Provided that $\varepsilon$ and $r$ are chosen small enough, one can
  assume that $2r + (\|u_0\| + r) \varepsilon \leqslant \alpha/3$.
  In particular, this implies $t_i > 2\alpha/3 > 0$.

  Using the integral form of the Mean Value Theorem, we get
  \begin{equation}
    \label{eq:w}
    w = e + \sum_{i=1}^k t_i \xi_i(\tilde{u}_0 +d)
    = e + \sum_{i=1}^k t_i \xi_i(\tilde{u}_0)
    + \int_0^1 \sum_{i=1}^k t_i \, \xi'_i(\tilde{u}_0+sd)[d] \,{\mathrm d} s .
  \end{equation}
  The third term can be rewritten as follows:
  \begin{align*}
 &\sum_{i=1}^k \langle{\tilde{u}_0}|{\xi_i(\tilde u_0)}\rangle \,
    \xi'_i(\tilde{u}_0)[d]
    + \sum_{i=1}^k
    \bigl(t_i - \langle{\tilde{u}_0}|{\xi_i(\tilde u_0)}\rangle \bigr)\,
    \xi'_i(\tilde{u}_0)[d] \\
 & + \int_0^1 \sum_{i=1}^k t_i \,
    \bigl(\xi'_i(\tilde{u}_0+sd)[d] - \xi'_i(\tilde{u}_0)[d]\bigr) \,{\mathrm d} s
    =: d_1 + d_2 + d_3 .
  \end{align*}
  Using Lemma~\ref{orthog-derivative} on $d_1$, one can write
  equation~\eqref{eq:w} as
  \begin{equation*}
    w = e + \sum_{i=1}^k
    \bigl(t_i - \langle{\tilde{u}_0}|{\xi_i'(\tilde{u}_0)[d]}\rangle \bigr)
    \, \xi_i(\tilde{u}_0)
    + d
    + d_2 + d_3 .
  \end{equation*}
  Since
  \[
    \bigl|\langle{\tilde{u}_0}|{\xi_i'(\tilde{u}_0)[d]}\rangle \bigr|
    \leqslant \|\tilde{u}_0\mathclose\|
    \mathopen\|\xi_i'(\tilde{u}_0)\mathclose\|
    \mathopen\|d\mathclose\|
    \leqslant (\|u_0\| + r) M r,
  \]
  we can assume $r$ was chosen small enough so that this is smaller
  that $\alpha/3$.  Recalling that $t_i > 2\alpha/3$, one sees that
  the coefficients of $\xi_i(\tilde{u}_0)$ are positive and therefore
  \begin{math}
    e + \sum  \bigl(t_i - \langle{\tilde{u}_0}|{\xi_i'(\tilde{u}_0)[d]}\rangle
   \bigr)    \, \xi_i(\tilde{u}_0)
    \in C_{\tilde u_0}
  \end{math}.

  The proof is complete if we show that~$d + d_2 + d_3 \in B(d,
  \delta \|d\|)$.  Using~\eqref{eq:t-i}, we deduce
  $|t_i| \leqslant
  \bigl|t_i - \langle{\tilde{u}_0}|{\xi_i(\tilde{u}_0)}\rangle\bigr|
  + \|\tilde u_0\| \leqslant 2r + (\|u_0\| + r) (\varepsilon + 1)$.
  %
  Thus the following estimates
  \begin{align*}
    \|d_2\|
    &\leqslant \sum_{i=1}^k
    \bigl|t_i - \langle{\tilde{u}_0}|{\xi_i(\tilde u_0)}\rangle\bigr|\,
    \|\xi'_i(\tilde{u}_0)\mathclose\|
    \mathopen\|d\|
    \leqslant k \bigl(2r + (\|u_0\| + r) \varepsilon\bigr) M \, \|d\|,
    \\
    \|d_3\|
    &\leqslant \sum_{i=1}^k |t_i|
    \sup_{s \in [{0,1}]}
    \|\xi'_i(\tilde{u}_0+sd) - \xi'_i(\tilde{u}_0)\mathclose\|
    \mathopen\|d\|
    \leqslant k \bigl(2r + (\|u_0\| + r) (\varepsilon + 1)\bigr)
    \varepsilon \|d\|,
  \end{align*}
  show that $\|d_i\| \leqslant \tfrac{1}{2}\delta \|d\|$, $i=2,3$,
  provided that the constants $\varepsilon$ and $r$ were chosen small
  enough.
\end{proof}

Lemma~\ref{convlemma2-xi} is the second key element
to prove the convergence up to a
subsequence.


\begin{lemma}  \label{convlemma2-xi}
  Let $\varphi$ be a peak selection for $(C_u)_{u \in{\mathcal A}}$ which verifies
  conditions~\eqref{Ac1} and \eqref{Ac-xi}.  Let
  $(u_n)_{n\in\mathbb{N}}$ and $(s_n)_{n\in\mathbb{N}}$ be given by the
  MPA (Algorithm~\ref{mpa}) with $s_n \in S(u_n)$ for all $n$. Let us assume
  that $\varphi$ is continuous, $\overline{\operatorname{Ran}\varphi} \subseteq {\mathcal A}$, and
  either
  \begin{subequations}
    \begin{gather}
      \label{eqsup}
      \exists \tau_1,\dots, \tau_k \in ({0,+\infty}),\
      \operatorname{dist} \Bigl(\Bigl\{\sum_{i=1}^k \tau_i \xi_i(u) :
      u\in\operatorname{Ran}\varphi \Bigr\}, \partial {\mathcal A}\Bigr) >0,
      \\
      \text{or }
      \label{eq:bdd-E}
      \begin{cases}
        \forall (v_n) \subseteq \operatorname{Ran}\varphi,\
        \bigl(\mathcal{E}(v_n)\bigr) \text{ is bounded from above}
        \Rightarrow
        (v_n) \text{ is bounded}
        \\
        \text{and }  \dim E < \infty.
      \end{cases}
    \end{gather}
  \end{subequations}
  If $\sum_{n=0}^{+\infty}s_n< +\infty$ then $(u_n)_{n\in\mathbb{N}}$
converges in ${\mathcal A}$.
\end{lemma}

\begin{proof}
  Let $k$ be given by the assumption~\eqref{Ac-xi}.
  For $i=1,\dots,k$, set
  $v_{i,n}:=\xi_i(u_n)$,
  and $d_n := - \frac{\nabla \mathcal{E}(u_n)}{
    \|\nabla\mathcal{E}(u_n)\|}$.
  Let $r$ be given by assumption~\eqref{Ac-xi}~(v) and $K_i$ be a
  bound for $\xi'_i$.  Denote $K := \max_{i=1,\dotsc, k} K_i$.

  By assumption~\eqref{Ac-xi} and as
  $\varphi(u_n + s_n d_n)\in \operatorname{int} C_{u_n + s_n d_n}$, we have
  \begin{equation*}
    v_{i,n+1} = \xi_i\bigl(\varphi(u_n+s_nd_n)\bigr)
    = \xi_i(u_n + s_n d_n)
  \end{equation*}
  for any $n\in\mathbb{N}$.
  Let $n^*$ be large enough so that for all $n \geqslant n^*,\ s_n < r$.
  Thus, for all $n \geqslant n^*$,
  \begin{equation}
    \label{eqProjector}
    \|v_{i,n+1}-v_{i,n}\|
    \leqslant K \|s_nd_{n}\|
    = K s_n.
  \end{equation}
  Since $\sum s_n < + \infty$, it follows that
  for any $i = 1,\dotsc,k$,
  $(v_{i,n})_{n\in\mathbb{N}}$ is a Cauchy sequence and therefore converges
  to, say, $v_{i,\infty}$.

  Let us assume~\eqref{eqsup} holds.
  Consider
  $\tilde{v}_n := \sum_{i=1}^k \tau_i v_{i,n}
  = \sum_{i=1}^k \tau_i \xi_i(u_n)$.  The sequence converges and
  its limit belongs to ${\mathcal A}$.
  Since $\varphi(\tilde{v}_n)=\varphi(u_n)=u_n$ and
  $\varphi$ is continuous, $(u_n)_{n\in\mathbb{N}}$ converges.
  Its limit lies in $\overline{\operatorname{Ran}\varphi}$ and thus in ${\mathcal A}$.

  If on the other hand \eqref{eq:bdd-E} holds, the fact that the
  sequence $(\mathcal{E}(u_n))$ is decreasing implies that $(u_n)$ is bounded.
  Let $(u'_n)$ be a subsequence of $(u_n)$.  Since $u'_n \in
  C_{u'_n}$, one can write $u'_n = e'_n + \sum_{i=1}^k t'_{i,n}
  \xi_i(u'_n)$ for some $e'_n\in E$ and $t'_{i,n} \in ({0,+\infty})$.
  As $(u'_n)$ is bounded, so are $(e'_n)$ and $|t'_{i,n}| =
  |\langle{u'_n}|{\xi_i(u'_n)}\rangle| \leqslant \|u'_n\|$.
  So, up to subsequences, $(e'_n)_{n\in\mathbb{N}}$ and
  $(t'_{i,n})_{n\in\mathbb{N}}$ converge to, say,
  $e'_\infty$ and $t'_{i,\infty}$.
  Thus $u'_n \to u'_\infty := e'_\infty + \sum t'_{i,\infty}
  v_{i,\infty}$.  Thanks to $\overline{\operatorname{Ran}\varphi} \subseteq {\mathcal A}$,
  $u'_\infty \in {\mathcal A}$.  But then the continuity of $\xi_i$ and $\varphi$ imply
  \begin{equation}
    \label{eq:lim-sum s_n}
    v_{i,\infty} = \xi_i(u'_\infty)
    \qquad\text{and}\qquad
    u'_n = \varphi(u'_n) \to \varphi(u'_\infty).
  \end{equation}
  If the same reasoning is performed with another subsequence
  $(u''_n)$, \eqref{eq:lim-sum s_n} implies that $\xi_i(u'_\infty) =
  \xi_i(u''_\infty)$ and therefore, in view
  of Definition~\ref{defpeak}~\eqref{phi-uniqueness}, $\varphi(u'_\infty) =
  \varphi(u''_\infty)$.  As the limit does not depend on the subsequence,
  the whole sequence $(u_n)$ converges in ${\mathcal A}$.
\end{proof}


\begin{remark} \rm
  If we wanted to
  seek sign-changing solutions using the cones $C_u := {\mathbb R}^+
  u^+ \oplus {\mathbb R}^+ u^-$ (as explained in the Introduction), then we
  would not be able to write~\eqref{eqProjector} in the above
  computation
  because the map $H^1_0 \to H^1_0 : u \mapsto u^+$ is not Lipschitz.
  This sheds some light on the difficulty of proving
  the convergence of the Modified Mountain Pass
  Algorithm~\cite{Neuberger97}.
\end{remark}

\begin{theorem}  \label{convtheorem}
  Assume $\varphi : {\mathcal A} \to {\mathcal A}$ is a continuous peak selection s.t.\
  $\overline{\operatorname{Ran}\varphi} \subseteq {\mathcal A}$ and
  the cones $(C_u)_{u \in {\mathcal A}}$ verify the
  conditions~\eqref{Ac1}, \eqref{Ac-xi} and~\eqref{eqsup} or~\eqref{eq:bdd-E}.
  Suppose moreover that $\mathcal{E} \in {\mathcal C}^1({\mathcal H}; {\mathbb R})$ satisfies the
  Palais-Smale condition in $\operatorname{Ran} \varphi$ and that $\inf_{u\in \operatorname{Ran} \varphi}
  \mathcal{E}(u)>-\infty$.  Then the sequence
  $(u_n)_{n\in \mathbb{N}}$ given by the  Mountain Pass
  Algorithm~\ref{mpa} possesses a  subsequence
  converging to a critical point of $\mathcal{E}$ in $\operatorname{Ran}\varphi$.
  In addition, all limit points  of
  $(u_n)_{n\in\mathbb{N}}$ are critical points of~$\mathcal{E}$.
\end{theorem}

\begin{proof}
  Let us start by showing that
$\bigl(\nabla\mathcal{E}(u_n)\bigr)_{n\in\mathbb{N}}$ converges
  to zero up to a subsequence. If not, we could assume there exist
  $\delta >0$ and $n_0\in\mathbb{N}$ such that, for any $n\geqslant n_0$,
  $\|\nabla \mathcal{E}(u_n)\| >\delta$. Then, for any $n\geqslant n_0$,
  the Deformation Lemma~\ref{deflemma} implies
  \begin{equation*}
    \mathcal{E}(u_{n+1})-\mathcal{E}(u_n) \leqslant -\alpha s_n\delta.
  \end{equation*}
  Thus, summing up,
  \begin{equation*}
    \lim_{n\to +\infty} \mathcal{E}(u_n)-\mathcal{E}(u_{n_0})
    = \sum_{n=n_0}^{+\infty} \mathcal{E}(u_{n+1})-\mathcal{E}(u_n)
    \leqslant -\delta \alpha \sum_{n=n_0}^{+\infty}s_n.
  \end{equation*}
  As the left-hand side is a real number ($\mathcal{E}$ is bounded from
  below on $\operatorname{Ran} \varphi$ and decreasing along
$(u_n)_{n\in\mathbb{N}}$), we have
  $\sum_{n=0}^{+\infty}s_n < +\infty$. So, by Lemma~\ref{convlemma2-xi},
  $u_n\to u^*\in {\mathcal A}$
  and $\|\nabla \mathcal{E}(u^*)\| \geqslant \delta$.
  By continuity of $\varphi$ at $u^* \in {\mathcal A}$,
  we obtain $\varphi(u^*)=u^*$ and, so,
  $u^*\in \operatorname{Ran} \varphi$. By Lemma~\ref{convlemma1}, there exists a
  neighborhood $V$ of $u^*$ and $s^*>0$ such that
  $S(u)\subseteq [{s^*,+\infty})$ for any $u\in V$.
  Consequently, there exists $n_0$
  such that, for any $n\geqslant n_0$, $s_n \geqslant s^*$, whence
  $\sum_{n=0}^{+\infty}s_n$ does not converge, which is a
  contradiction.

  In conclusion, there exists a subsequence $(u_{n_k})_{k\in\mathbb{N}}$ of
  $(u_n)_{n\in\mathbb{N}}$ s.t.\ $\|\nabla \mathcal{E}(u_{n_k})\| \to 0$
  when $k\to +\infty$. As $\mathcal{E}$ satisfies the Palais-Smale
  condition, $(u_{n_k})_{k\in\mathbb{N}}$ possesses a subsequence converging
  to a critical point of $\mathcal{E}$.

  Concerning the second statement of the theorem, the argument is very
  similar. Let $(u_{n_k})_{k\in\mathbb{N}}$ be a convergent subsequence and
  assume on the contrary that $u := \lim_{k\to \infty}u_{n_k}$ is not a
  critical point of $\mathcal{E}$. In that case, on one hand, there exists
  $\delta >0$ and $k_1\in\mathbb{N}$ such that, for any $k\geqslant k_1$,
  $\|\nabla \mathcal{E}(u_{n_k})\| \geqslant \delta$.  By Lemma~\ref{deflemma}, we
  have
  \begin{equation*}
    \forall k\geqslant k_1,\quad
    \mathcal{E}(u_{n_{k+1}})-\mathcal{E}(u_{n_k}) \leqslant -\alpha\delta s_{n_k}.
  \end{equation*}
  On the other hand, as $u\in \operatorname{Ran} \varphi$, we have by
 Lemma~\ref{convlemma1} that
  \begin{equation*}
    \exists\ s^*>0,\ \exists k_2\in\mathbb{N},\quad
 \forall k\geqslant k_2,\ s_n\geqslant s^*.
  \end{equation*}
  So, for large $k$, $\mathcal{E}(u_{n_{k+1}})-\mathcal{E}(u_{n_k})\leqslant
  -\frac{\alpha}{2}\delta s^*$, which is a contradiction because
  $\bigl(\mathcal{E} (u_n)\bigr)_{n\in\mathbb{N}}$ is a convergent sequence.
\end{proof}

\begin{remark} \rm
 By previous Remarks~\ref{uniform1} and~\ref{uniform2}, we conclude
that we could get the convergence up to a subsequence using the
equation~\eqref{step} only at $u_0$. Thus, the uniform form of
Lemma~\ref{deflemma} is not required (and we could
consider that $\varphi(u)$ is a local maximum of $\varphi$ on
$C_u$ instead of a global maximum). Nevertheless,  we have kept the
uniform setting along the paper as it will be required in
Section~\ref{convmpa2}.
\end{remark}

The following special case of~\eqref{Ac-xi} is important for the
applications.
\begin{equation}
%  \tag{$AC_{4}$}
\label{Ac4}
  \left.\begin{minipage}{0.83\linewidth}
      There exist a closed subspace $E\subseteq {\mathcal H}$ (possibly
      infinite dimensional) and
      linear continuous projectors $P_i :{\mathcal H}\to E^{\perp}$,
      $i=1,\dots,k$, for some $k\in\mathbb{N}$,  such that
      \begin{itemize}
      %\item $P_i(e)=0$ for all $e\in E$ and $i=1,\dots,k$;
      \item $\forall\ u\in {\mathcal H}$, $P_i(u)\perp P_j(u)$ whenever $i\ne j$;
      \item $E \oplus \sum_{i=1}^k \operatorname{Ran} P_i = {\mathcal H}$.
      \end{itemize}
      For all $u \in {\mathcal H}$, set $C_u := E
      \oplus \bigl\{ \sum  t_i P_i(u) \bigm| t_i\geqslant 0 \text{ for all }
      i\bigr\}$.
    \end{minipage}
  \ \right\}
\end{equation}

Let us now sketch the proof that~\eqref{Ac4} implies both~\eqref{Ac1}
and~\eqref{Ac-xi}.
Consider
${\mathcal A} := \{ u \in {\mathcal H} \mid P_i(u) \ne 0 \text{ for all } i \}$ and
$\xi_i(u) := \frac{P_i(u)}{\|P_i(u)\|}$.
Clearly $\xi_1, \dotsc, \xi_k$ are ${\mathcal C}^1$ functions on ${\mathcal A}$.
Moreover $e + \sum t_i P_i(u) \in \operatorname{int} C_u$ if and only if
all $t_i > 0$.  Since $u = P_E(u) + \sum_{i=1}^k P_i(u)$ where $P_E$
denotes the orthogonal projection onto $E$, one has $u \in
\operatorname{int} C_u$.
Given the definitions of ${\mathcal A}$ and $\xi_i$, points (i), (iii) and~(iv)
of~\eqref{Ac-xi} are straightforward.
A simple computation shows that
$\xi'_i(u)\bigl[\sum t_j P_j(u)\bigr]$ is a multiple of $P_i(u)$
whence~(ii) follows.
Finally, as $\|\xi'_i(u)\| = O(1/\|{P_i(u)}\|)$, (v) will hold
provided $\|P_i(u)\|$ is bounded away from $0$ when $u \in \operatorname{Ran}\varphi$.
Note that this latter condition also ensures that
$\overline{\operatorname{Ran}\varphi} \subseteq {\mathcal A}$.
We remark that these cones satisfy property \eqref{eqsup} with $\tau_1 =
\cdots = \tau_k = 1$ because $\operatorname{dist}(\sum \xi_i(u), \partial{\mathcal A})
= \min_j \|P_j(\sum \xi_i(u))\| = 1$.

Thus, as a corollary of Theorem~\ref{convtheorem}, we get the
following proposition.
It can be thought as an abstract version of the convergence
results in~\cite{sym1,zhou1,zhou2}.

\begin{proposition}
  \label{conv-P_i}
  Let us consider $\varphi : {\mathcal A} \to {\mathcal A}$ a continuous
  peak selection with the cones $(C_u)_{u \in {\mathcal A}}$ being given by
  condition~\eqref{Ac4} and ${\mathcal A} := \{u\in {\mathcal H} \mid
  P_i(u)\ne 0 \text{ for all } i=1,\dots,k\}$.
  Assume that
  $\inf_{u\in \operatorname{Ran} \varphi} \|P_i (u)\| >0$ for all
  $i=1,\dots,k$, that $\mathcal{E} \in {\mathcal C}^1({\mathcal H}; {\mathbb R})$ satisfies the
  Palais-Smale condition in $\operatorname{Ran} \varphi$ and that $\inf_{u\in \operatorname{Ran} \varphi}
  \mathcal{E}(u)>-\infty$.   Then the sequence
  $(u_n)_{n\in \mathbb{N}}$ given by the  Mountain Pass
  Algorithm~\ref{mpa} possesses a  subsequence
  converging to a critical point of $\mathcal{E}$ in $\operatorname{Ran}\varphi$.
  In addition, all limit points  of
  $(u_n)_{n\in\mathbb{N}}$ are critical points of~$\mathcal{E}$.
\end{proposition}


In Theorem~\ref{convtheorem}, the Palais-Smale condition is required. For
the particular  case of  ${\mathcal H} = H^1({\mathbb R}^N)$,  this
condition does not generally hold, as ${\mathbb R}^N$ is not bounded, some
mass of Palais-Smale sequences may be lost at infinity i.e.,
\[
\lim_{R\to\infty} \limsup_{n \to\infty} \int_{{\mathbb R}^N \setminus B(0,R)}
u_n^2 > 0.
\] Fortunately,
$H^1({\mathbb R}^N)$ enjoys the following compactness condition (see
for example the
 paper~\cite{lieb} for a proof): for any
bounded sequence $(u_n)_{n\in\mathbb{N}}\subseteq H^{1}({\mathbb R}^N)$
staying away from zero, there exists
$(x_n)_{n\in\mathbb{N}}\subseteq {\mathbb Z}^N$ such that the sequence
$\bigl(u_n(\cdot + x_n)\bigr)$, where $u_n(\cdot + x_n)$
denotes the function $x \mapsto u_n(x + x_n)$, converges weakly,
up to a subsequence, to a non-zero function.  This is
enough to get the convergence up to a subsequence.

\begin{proposition}  \label{convinrn}
  Assume the hypotheses of Theorem~\ref{convtheorem} hold, except for
  the Palais-Smale condition.
  Let ${\mathcal H} := H^1({\mathbb R}^N)$ and $(u_n)_{n\in \mathbb{N}}$
be the sequence given  by the Mountain Pass Algorithm~\ref{mpa}.
  If, for any $u\in {\mathcal H}$ and
  $x\in {\mathbb Z}^N$, $\mathcal{E}\bigl(u(\cdot +x)\bigr) = \mathcal{E}(u)$
  and if $ \nabla \mathcal{E} : {\mathcal H} \to {\mathcal H}$ is continuous for the
  weak topology on ${\mathcal H}$,
  then there exists a sequence $(x_n)_{n\in\mathbb{N}}\subseteq
  {\mathbb Z}^N$ such that $\bigl(u_n(\cdot + x_n)\bigr)_{n\in\mathbb{N}}$ converges weakly up
  to a subsequence to a nontrivial critical point of $\mathcal{E}$.
 \end{proposition}

\begin{proof}
  We will only briefly sketch the proof.
  As $(u_n)_{n\in\mathbb{N}}$ is bounded in $H^1({\mathbb R}^N)$
 and stays away from
  $0$, the compactness condition recalled above implies
  that there exists a sequence $(x_n)_{n\in\mathbb{N}}\subseteq {\mathbb Z}^N$
 such  that $u(\cdot + x_n)$ converges weakly, up to a subsequence, to
  $u^*\neq 0$. Intuitively, the translations ``bring back'' some mass
  that $u_n$ may loose at infinity.

  Using the translation invariance of
  $\mathcal{E}$, the corresponding equivariance of $\nabla\mathcal{E}$ and
  the weak continuity of $\nabla\mathcal{E}$,
  we conclude that $u^*$ is a critical
  point of $\mathcal{E}$.
\end{proof}


\subsection{Convergence of the whole sequence}
\label{convmpa2}

In this section, we refine the stepsize used previously to get the
convergence of the whole sequence generated by
Algorithm~\ref{mpa}. We require that the stepsize $s_n \in
\tilde{S}(u_0) := \tilde{S}^{*}(u_0)\cap \big(\frac{1}{2}\sup
  \tilde{S}^{*}(u_0), +\infty \big)$ where
  \begin{align*}
    \tilde{S}^{*}(u_0)
    := \Bigl\{ s_0>0  :&  \forall\ s \in ({0, s_0}],\
     u_s := u_0 -
    s\frac{\nabla\mathcal{E}(u_0)}{
      \|\nabla\mathcal{E}(u_0)\|} \in {\mathcal A}
    \text{ and }\\
    &\mathcal{E}\bigl(\varphi(u_s)\bigr) - \mathcal{E}(u_0)
    < -\alpha s \mathopen\|\nabla \mathcal{E}(u_0)\|   \Bigr\}  .
  \end{align*}
Using the Deformation Lemma~\ref{deflemma}, we get that $\tilde S(u_n) \ne
\varnothing$ as long as $u_n$ is not a critical point.
Moreover, working as previously, we get
results~\ref{convlemma1}, \ref{convlemma2-xi} and \ref{convtheorem} for
this new choice of stepsizes. Let
us remark that, this time,  we really need that
inequality~\eqref{step} is valid in a neighborhood of $u_0$ to get
Lemma~\ref{convlemma1}.
This new stepsize will allow us  to
control the energy for any $0<s\leqslant s_0$.
Under a ``localization'' assumption, we now prove that
the whole sequence $(u_n)_{n\in\mathbb{N}}$ given by the
Mountain Pass Algorithm~\ref{mpa}
converges to a nontrivial critical point of $\mathcal{E}$.

\begin{theorem}   \label{convthm}
  Assume that $u$ is the unique critical point of $\mathcal{E}$ in the ball $B(u,
  \delta)$ for some $\delta >0$.
  Under the same assumptions as those of Theorem~\ref{convtheorem}, if there
  exists $n^*\in \mathbb{N}$ such that $\mathcal{E}(u_{n^*})<a:= \inf_{v\in
    \partial B(u,\delta)\cap \operatorname{Ran} \varphi}\mathcal{E}(v)$ and $u_{n^*}\in B(u,\delta)$
  then the  sequence $(u_n)_{n\in\mathbb{N}}$ produced by
  Algorithm~\ref{mpa} with stepsizes $s_n \in \tilde{S}(u_n)$
  converges to $u$.
\end{theorem}

\begin{proof}
  For any $m> n^*$, we claim that $u_m \in B(u, \delta)$. If not,
  as $u_{n^*}\in B(u,\delta)$,
  there exists $m\geqslant n^*$ such that $u_m \in B(u, \delta)$ and
  $u_{m+1}= \varphi\bigl(u_m -s_m\frac{\nabla \mathcal{E}(u_m)}{
    \|\nabla \mathcal{E}(u_m)\|} \bigr) \notin
  B(u, \delta)$, with $s_m\in \tilde{S}(u_m)$.
  %
  By continuity, there exists $0<s\leqslant s_m$ such that
  $\varphi\bigl(u_m -s\frac{\nabla \mathcal{E}(u_m)}
  {\|\nabla \mathcal{E}(u_m)\|} \bigr)\in \partial B(u, \delta)\cap \operatorname{Ran} \varphi$.
  This is a contradiction
  because, by the definition of $s_m$ and as $\mathcal{E}$ is
  decreasing along $(u_n)_{n\in\mathbb{N}}$,  we have
  $a\leqslant \mathcal{E}\bigl(\varphi(u_m -s\frac{\nabla \mathcal{E}(u_m)}{
    \|\nabla
      \mathcal{E}(u_m)\|} )\bigr)\leqslant \mathcal{E}(u_m)\leqslant  \mathcal{E}(u_{n^*})<a$.

  As $u$ is the unique critical point in $B(u,\delta)$, by
  Theorem~\ref{convtheorem}, $u$ is the unique accumulation point of
  $(u_n)_{n\in\mathbb{N}}$. So, $u_n$ converges to $u$.
\end{proof}

\section{Applications}
\label{applmpa}

\subsection{Application to indefinite problems}

For problem~\eqref{eq:Szulkin},
the energy functional $\mathcal{E}$ given by~\eqref{functional-indefinite}
is defined on ${\mathcal H}:=
H^1_0(\Omega)$. Let us denote the decomposition ${\mathcal H}= {\mathcal H}^{(-)} \oplus {\mathcal H}^{(+)}$
corresponding to the spectral decomposition of
$-\Delta +V$ with respect to the positive and negative part of the
spectrum.  For any $u$, we let $u^{(-)}\in {\mathcal H}^{(-)}$ and $u^{(+)}\in
{\mathcal H}^{(+)}$ be the unique elements
such that $u = u^{(-)}+u^{(+)}$.
Let us remark that the case ${\mathcal H}^{(-)} = \{0\}$ corresponds the
traditional Mountain Pass Algorithm with a positive definite linear operator.

We choose the following peak selection.
Let ${\mathcal A} := {\mathcal H} \setminus {\mathcal H}^-$
and, for any $u\in {\mathcal A}$, let $C_u$ be the cone
$C_u := {\mathcal H}^{(-)} \oplus {\mathbb R}^+ u = {\mathcal H}^{(-)} \oplus {\mathbb R}^+ u^{(+)}$.
The peak selection $\varphi$ for $(C_u)_{u\in {\mathcal A}}$ is the map
\begin{equation*}
  \varphi: {\mathcal A}\to {\mathcal A} : u\mapsto \varphi(u)
\end{equation*}
such that, for all $u \in {\mathcal A}$,
$\varphi(u)$ maximizes $\mathcal{E}$ on $C_u$.   To prove that $\varphi$ is
continuous, we refer to the original paper~\cite{szulkin}.
Is is easy to check that these cones verify
\eqref{Ac4}.  Indeed it suffices to consider $E=
{\mathcal H}^{(-)}$, $k=1$ and $P_1 : {\mathcal H}\to E^\perp$, the orthogonal
projection onto~$E^\perp$.

To apply Proposition~\ref{conv-P_i},  we need to verify the
following assumptions  on $\mathcal{E}$: on a bounded domain $\Omega$,
\begin{enumerate}
\item[(i)] $0$ does not belong to
  $\overline{\operatorname{Ran}(P_1 \circ \varphi)}$:
  it comes from the
  fact that $0$ is a strict local minimum of $\mathcal{E}$ on $E^{\perp}={\mathcal H}^{(+)}$
  (see~\cite{szulkin});
\item[(ii)] $\mathcal{E}\in{\mathcal C}^1\bigl(H^1_0(\Omega); {\mathbb R}\bigr)$: it follows from
  standard arguments;
\item[(iii)] $\mathcal{E}$ verifies the Palais-Smale condition on $\operatorname{Ran} \varphi$:
  see~\cite{szulkin};
\item[(iv)] $\inf_{u\in \operatorname{Ran} \varphi}\mathcal{E}(u)>-\infty$: actually $\mathcal{E}$ is bounded from
  below by $0$ on $\operatorname{Ran} \varphi$,
  see~\cite{szulkin}.
\end{enumerate}
In conclusion, Proposition~\ref{conv-P_i} applies and gives
the convergence up to a
subsequence of the sequence $(u_n)$ generated by
Mountain Pass Algorithm~\ref{mpa} for this indefinite
problem provided that the domain $\Omega$ is bounded.

Let us now sketch what happens about the convergence up to a
subsequence when $\Omega = {\mathbb R}^N$.
As $(\mathcal{E}(u_n))_{n\in\mathbb{N}}$ is
decreasing (see Proposition~\ref{stability}) and is bounded away from zero, we
have that $(u_n)_{n\in\mathbb{N}}$ is  bounded and stays away from zero in
$H^1({\mathbb R}^N)$ (see~\cite{szulkin}).  On the other hand, $V$ is assumed
to be $1$-periodic, thus $\mathcal{E}\bigl(u(\cdot + x)\bigr) = \mathcal{E}(u)$
for any $u \in {\mathcal H}$ and $x \in {\mathbb Z}^N$.
It is not difficult to check
that $\nabla\mathcal{E}: {\mathcal H} \to {\mathcal H}$ is continuous for
the weak topology on~${\mathcal H}$.  Thus, Theorem~\ref{convinrn}
asserts that, if $(u_n)$ is the sequence generated by the MPA,
there exists a sequence of translations $(x_n) \subseteq {\mathbb Z}^N$
such that $(u_n(\cdot + x_n))_{n\in\mathbb{N}}$
converges weakly, up to a subsequence, to a nontrivial critical point
$u^*$ of $\mathcal{E}$.
Moreover, if $\mathcal{E}(u_n)
\to \inf_{u\in\operatorname{Ran} \varphi}\mathcal{E}(u)$, then it can be proved
that the above convergence is strong.  The idea is that,
if it does not converge strongly, some mass is lost at
infinity.  At the limit, this mass will take away a quantity of
energy greater or equal to
$\inf_{u \in \operatorname{Ran}\varphi} \mathcal{E}(u) > 0$, a contradiction.


\paragraph{Numerical experiments}
Let us start by giving some details on the computation of various
objects used in the MPA.
Functions in ${\mathcal H}$ will be approximated using
$P^1$-finite elements on a Delaunay triangulation of $\Omega$
generated by the software Triangle~\cite{Triangle}.
The matrix of the quadratic
form $(u_1, u_2) \mapsto \int_{\Omega} \nabla u_1 \nabla u_2$ is
readily evaluated on the finite elements basis.
For $(u_1, u_2) \mapsto \int_{\Omega} V(x) u_1 u_2 \,{\mathrm d} x$
and the various integrals involving $u$ to a power,  a quadratic
integration formula on each triangle is used.
The gradient $g := \nabla\mathcal{E}(v)$ is computed in the usual way:
the function $g \in {\mathcal H}$ is the solution of the linear system of equations
$\forall \varphi \in {\mathcal H}$, $(g | \varphi)_{{\mathcal H}} =
\mathcal{E}'(v)[\varphi]$.
%
In practice,  the peak selection  $\varphi$ must be evaluated with great
accuracy to obtain satisfying
results.  For this, we use a
limited-memory quasi-Newton code for bound-constrained
optimization~\cite{L-BFGS-B}.
%
The program stops when the gradient
of the energy functional at the approximation has a norm less than
$10^{-4}$.

As an illustration, we consider $\Omega = ({0,1})^2$,
$V \in{\mathbb R}$ constant and $p=4$.  Let us remark that ${\mathcal H}^{(-)}$
is then formed by eigenfunctions of $-\Delta +V$ with negative eigenvalues.
In dimension~$2$, the eigenvalues $\lambda_i$ of $-\Delta$ on the
square $({0,1})^2$ with zero Dirichlet boundary conditions
are given
by $\pi^2 (n^2 + m^2)$ with $n,m = 1,2,\dots$\@ The
related eigenfunctions are given by $\sin (n\pi x )\sin(m\pi y)$.
We get $\lambda_1 = 2\pi^2 \approx 19.76$,
$\lambda_2 = \lambda_3 = 5\pi^2 \approx 49.48$ (a double
eigenvalue),
$\lambda_4 = 8\pi^2 \approx 78.95$,
$\lambda_5 = \lambda_6 = 10\pi^2 \approx 98.69$,...

Figure~\ref{carrelam} depicts four
non-zero solutions  approximated by the Algorithm~\ref{mpa} for four
different values of $V$.  The algorithm was always started from
$u_0(x,y) := xy(x-1)(y-1)$.
The graphs on the left-hand side are
given for the values $V = 0$ ($\dim {\mathcal H}^{(-)} = 0$) and
$-\lambda_2 < V = -21 < -\lambda_1$ ($\dim {\mathcal H}^{(-)}=1$). The
graphs on the right-hand side are given for $-\lambda_4 < V = -50
< -\lambda_3$ ($\dim{\mathcal H}^{(-)} = 3$) and $-\lambda_5 < V = -80
< -\lambda_4 $ ($\dim {\mathcal H}^{(-)} =4$). In Table~\ref{tablelam}, we
present some characteristics of the solutions.

\begin{figure}[htb]
 \begin{center}
\begin{tabular}{@{}c@{}c@{}}
  \includegraphics[width=0.45\linewidth]{fig1a} % lambda0
& \includegraphics[width=0.45\linewidth]{fig1b} % u2 
\\ $V=0$ & $V=-50$  
\\
  \includegraphics[width=0.45\linewidth]{fig1c} % u1
& \includegraphics[width=0.45\linewidth]{fig1d} % u3
\\ $V=-21$ & $V=-80$ 
\end{tabular}
\end{center}
\caption{MPA solutions for an indefinite problem on a square}
  \label{carrelam}
\end{figure}

\begin{table}[ht]
  \begin{center}
    \begin{tabular}{|rccc|}\hline
$V$ & $\|\nabla \mathcal{E}\|$ & \# of steps  & $\mathcal{E}(u)$\\
      \hline $0$ &   $6.0\times 10^{-5}$ & 7  & 37.89\\
      \hline $-21$ & $6.4\times 10^{-5}$ & 48 & 70.43\\
      \hline $-50$ & $5.3\times 10^{-5}$ & 113& 91.42\\
      \hline $-80$ & $6.5\times 10^{-5}$ & 44 & 35.06 \\ \hline
    \end{tabular}
  \end{center}
\vspace{2mm}
  \caption{Characteristics of approximate solutions to an indefinite problem.}
  \label{tablelam}
\end{table}

For $V = 0$, we remark that the approximation is even w.r.t.\
any symmetry of the square
and is positive. It was expected and it is actually already known in this case
(i.e.\  for the problem $-\Delta u = |u|^{p-2}u$) that
ground state solutions have the same symmetries
as the first eigenfunctions of $-\Delta$ (see~\cite{gnn,bbgv}).

For $V = -21$, the approximation has two nodal domains and a
diagonal as nodal line. It seems to respect the
symmetries of a second eigenfunction of $-\Delta$, which can be explained
as follows.  When $V = 0$, it is
proved \cite{bbgv} that, for $p$ close to $2$, least energy nodal solutions
have the same symmetries as
their projections onto the second eigenspace of $-\Delta$. On the
square, it is even conjectured that the projection  must be a
function odd w.r.t.\ a diagonal.
In view of the bifurcation diagrams computed by
J.~M.~Neuberger~\cite{Neuberger-GNGA, Neuberger-Newton},
the least energy nodal solution for $V \in ({-\lambda_1, 0}]$
becomes the solution with lowest energy when
$V \in ({-\lambda_2, -\lambda_1}]$ and no bifurcation happens
along the way.  So it is reasonable (and this is supported by
the bifurcation diagrams) that they keep the same symmetries
along the whole branch.

We also observe that, for $V = -50$  (resp.\ $-80$), the
approximation  seems to respect the
symmetries of (and has the ``same form'' as) a fourth (resp.\ fifth)
eigenfunction of $-\Delta$. Their
number of bumps corresponds to their Morse index
($\dim{\mathcal H}^{(-)} + 1$).

All those considerations support the conjecture that
if $-\lambda_{n}< V < -\lambda_{n-1}$ then, at least for $p$
  small enough, ground state solutions respect the symmetries of a
  $n^{\text{th}}$ eigenfunction of~$-\Delta$.


\subsection{Application to systems}

In this section we will perform numerical experiments for the
system~\eqref{eqNoris-general}.
The corresponding energy
functional~\eqref{functional-system} is defined on
${\mathcal H}= H^1_0(\Omega,{\mathbb R}^k)$ endowed with the norm $\|u\|^2 =
\int_{\Omega} |\nabla u|^2 = \sum_i \int_{\Omega} |\nabla
  u_i|^2\,{\mathrm d} x$.  In \cite{noris}, B.~Noris and G.~Verzini
prove that the minimization of $\mathcal{E}$ on
\begin{equation*}
  {\mathcal N} := \Bigl\{u\in {\mathcal A} :
  \forall i=1,\dots,k,\ \int_{\Omega}
  |\nabla u_i|^2\,{\mathrm d} x
  = \int_{\Omega} \partial_i F(u)u_i\,{\mathrm d} x
  \Bigr\},
\end{equation*}
where
${\mathcal A}  := \{u\in {\mathcal H} \mid u_i\neq 0 \text{ for every } i\}$,
yields a solution $u=(u_1,\dots, u_k) = \sum u_i e_i$
with $u_i \neq 0$ for all $i=1,\dots,k$ provided that the following
assumptions are satisfied:
there exist  $p\in ({2,2^*})$, $C_F >0$ and
$\delta >0$ such that, for any $u,\lambda \in {\mathbb R}^k$, one has
\begin{enumerate}
\item[(i)] \label{F1} $\sum_{i,j}|\partial^2_{i,j} F(u)| \leqslant C_F |u|^{p-2}$,
  $\sum_{i} |\partial_i F(u)| \leqslant C_F |u|^{p-1}$ and
  $|F(u)|\leqslant C_F |u|^p$;
\item[(ii)]  $\sum_{i,j} \partial^2_{i,j} F(u)\lambda_i u_i \lambda_j u_j -
  (1+\delta) \sum_{i}\partial_i F(u)\lambda_i^2 u_i\geqslant 0$;
\item[(iii)] for every $i$ there exists $\bar{u}_i >0$ such that $\partial_i
  F(\bar{u}_i e_i)>0$;
\item[(iv)] \label{F4}
  $\partial_i F(u)u_i \leqslant \partial_i F(u_ie_i)u_i$ for every $i$.
\end{enumerate}
The first three assumptions are traditional in the framework of
variational methods. The last one insures $u_i\ne 0$ for all $i$.
In this section, we will use the Mountain Pass
Algorithm~\ref{mpa} with the following peak selection.
For any $u= (u_1,\dots, u_k)\in {\mathcal A}$, we consider the
cone $C_u := \{(t_1 u_1,\dots, t_k u_k) \mid t_i\geqslant 0 \text{ for all }
i=1,\dots,k\}$.
The peak selection $\varphi$ for $(C_u)_{u\in{\mathcal A}}$ is the map
\begin{equation*}
  \varphi: {\mathcal A}\to {\mathcal A} : u\mapsto \varphi(u)
\end{equation*}
such that $\varphi(u)$ maximizes $\mathcal{E}$ on $C_u$. Under the additional
hypothesis that $\sum_i \partial_i F(u) u_i \geqslant 0$, the second
assumption plays the role of the traditional super-quadraticity and
implies that $\varphi$ is well-defined as a peak selection. In
fact, if $u\in {\mathcal A}$
verifies $\mathcal{E}'(u)[(\lambda_1
u_1,\dots, \lambda_k u_k)] =0$ for any $(\lambda_1,\dots,\lambda_k)
\in {\mathbb R}^k$
then   $u$ is a strict local maximum of
$\mathcal{E}$ on $C_u$. It implies the uniqueness of the global maximum
of $\mathcal{E}$ on $C_u$. Moreover, $\varphi$ is continuous.

To see that assumption~\eqref{Ac4} is satisfied, it suffices to
take $E = \{0\}$ and, for
$i=1,\dots,k$, $P_i(u) = P_i\bigl((u_1,\dots, u_k)\bigr) := u_i e_i$,
i.e., $P_i$ is the projection onto the $i^{\text{th}}$ component of $u$.
Finally,
let us quickly run through the assumptions of Proposition~\ref{conv-P_i}:
\begin{enumerate}
\item[(i)] $\operatorname{dist} (\operatorname{Ran}\varphi,\partial{\mathcal A})>0$: see~\cite{noris};
\item[(ii)] $\mathcal{E}\in{\mathcal C}^1({\mathcal H}; {\mathbb R})$: it follows from standard arguments;
\item[(iii)] $\mathcal{E}$ verifies the Palais-Smale condition on $\operatorname{Ran} \varphi$:
  see~\cite{noris};
\item[(iv)] $\inf_{u\in \operatorname{Ran} \varphi}\mathcal{E}(u)>-\infty$: actually $\mathcal{E}$ is bounded from
  below by $0$ on $\operatorname{Ran} \varphi$ (see~\cite{noris}).
\end{enumerate}
In conclusion, Proposition~\ref{conv-P_i} applies and gives
the convergence, up to a
subsequence, of the sequence $(u_n)$ generated by the
Mountain Pass Algorithm~\ref{mpa}.


\subsection*{Numerical experiments}

For the numerical experiments, we will consider the following
particular case of equation~\eqref{eqNoris-general}:
\begin{equation}  \label{eqNoris}
    \begin{gathered}
      -\Delta u_i(x)
      = \mu_i u_i^3 + u_i\sum_{j\ne i} \beta_{i,j}u_j^2 ,\quad x\in\Omega, \\
      u_i(x)=0,\quad x \in\partial\Omega ,
    \end{gathered}
   \qquad  i=1,\dots,k,
\end{equation}
where $\beta_{i,j} = \beta_{j,i}$ and $\Omega$ is a bounded
domain of~${\mathbb R}^2$.
This system is modeling a competition between $k$
populations.  We will focus on the case $\Omega = ({0,1})^2$
and $k=2$.  In this setting, the assumptions
(\ref{F1})--(\ref{F4}) stated above boil down to
\begin{equation}
  \label{eq:sys-beta-cond}
  \mu_1 > 0,\quad
  \mu_2 > 0,\quad\text{and}\quad
  -\sqrt{\mu_1 \mu_2} \leqslant \beta_{1,2} \leqslant 0.
\end{equation}
Let us remark that  the condition $\sum_i \partial_i F(u) u_i \geqslant 0$
discussed in the previous section is also verified in this range.

Let us now give the outcome of the algorithm for various choices of
$(\mu_1, \mu_2, \beta_{1,2})$.  The MPA will always start with the
function $u_0 = (u_{0,1}, u_{0,2}) \in {\mathcal A}$ where
$u_{0,1}(x,y) =  u_{0,2}(x,y) = xy(1-x)(1-y)$ and stops when the
norm of the gradient is less than~$10^{-4}$.

First we choose $(\mu_1, \mu_2, \beta_{1,2}) = (1, 4, -1)$.  The
numerical solution $(u_1, u_2)$ is depicted on 
Figure~\ref{carre-sys-1} and some characteristics are given in
Table~\ref{table-sys}.  In this case, the assumptions
\eqref{eq:sys-beta-cond} are satisfied so the algorithm
converges to a solution $(u_1, u_2)$ with $u_1 > 0$ and $u_2 > 0$ as
expected.
Notice also that the solutions $u_1$ and $u_2$ are even w.r.t.\
axes of symmetry of the square.

As a second choice, we consider $(\mu_1, \mu_2, \beta_{1,2}) = (1, 4,
0.5)$.  The MPA solution $(u_1, u_2)$ is depicted on
Figure~\ref{carre-sys-0.5} and some characteristics are given in the
 second row of Table~\ref{table-sys}.  Despite the fact that the assumptions
\eqref{eq:sys-beta-cond} are not satisfied anymore, the solution is
similar to that found in the first case.  If we enlarge $\beta_{1,2}$
further and choose $(\mu_1, \mu_2, \beta_{1,2}) = (1, 4, 1.2)$, the
algorithm still converges (see the third row of Table~\ref{table-sys}),
but this time, the second component vanishes
(see Figure~\ref{carre-sys-1.2}).  What happens is that, at the very
first step, $u_2 = 0$ and then the MPA essentially proceeds as if the
system was only consisting in the first equation.

\begin{figure}[htb]
\vspace{-20mm}
  \begin{center}
      \includegraphics[width=0.45\linewidth, clip=true ]{fig2a} % sys_m1_1.pdf
\quad \includegraphics[width=0.45\linewidth, clip=true ]{fig2b} % sys_m1_2.pdf
  \end{center}
\vspace{-20mm}
\caption{MPA solution for the system with
    $(\mu_1, \mu_2, \beta_{1,2}) = (1, 4, -1)$.}
  \label{carre-sys-1}
\end{figure}

\begin{figure}[htb]
\vspace{-20mm}
  \begin{center}
   \includegraphics[width=0.45\linewidth]{fig3a} % sys_05_1.pdf
    \hfil
\quad  \includegraphics[width=0.45\linewidth]{fig3b} % sys_05_2.pdf
  \end{center}
\vspace{-20mm}
  \caption{MPA solution for the system with
    $(\mu_1, \mu_2, \beta_{1,2}) = (1, 4, 0.5)$}
  \label{carre-sys-0.5}
\end{figure}

\begin{figure}[htb]
\vspace{-20mm}
  \begin{center}
  \includegraphics[width=0.45\linewidth]{fig4a} % sys_12_1.pdf
\quad  \includegraphics[width=0.45\linewidth]{fig4b} % sys_12_2.pdf
  \end{center}
\vspace{-20mm}
\caption{MPA solution for the system with
    $(\mu_1, \mu_2, \beta_{1,2}) = (1, 4, 1.2)$.}
  \label{carre-sys-1.2}
\end{figure}

\begin{table}[ht]
  \begin{center}
    \begin{tabular}{|cccccc|} \hline
       & $\|\nabla \mathcal{E}(u)\|$
      &\# steps & $\mathcal{E}(u)$
      &$\max u_1$  &$\max u_2$\\
      \hline
      $(1, 4, -1) $& $7.9\times 10^{-5}$ & 11& 88.4&  8.6&  5.4\\ \hline
      $(1, 4, 0.5)$& $5.4\times 10^{-5}$ & 11& 40.4&  6.4&  2.4\\ \hline
      $(1, 4, 1.2)$& $5.2\times 10^{-5}$ & 11& 39.9&  6.6&  0.0\\ \hline
    \end{tabular}
  \end{center}
\vspace{2mm}
  \caption{Characteristics of the solution to system~\eqref{eqNoris}.}
  \label{table-sys}
\end{table}

\subsection*{Acknowledgments}
 The authors are partially supported by a grant from the
 National Bank of Belgium and by the program
``Qualitative study of solutions of variational elliptic partial differerential
  equations. Symmetries, bifurcations, singularities, multiplicity and numerics''
  of the FNRS, project 2.4.550.10.F of the Fonds de la Recherche
  Fondamentale Collective.

C. Grumiau would like to dedicate this work to his grandmother who
  passed away a few weeks ago.

\begin{thebibliography}{10}

\bibitem{bbgv}
Denis Bonheure, Vincent Bouchez, Christopher Grumiau, and Jean Van~Schaftingen.
\newblock Asymptotics and symmetries of least energy nodal solutions of
  {L}ane-{E}mden problems with slow growth.
\newblock {\em Commun. Contemp. Math.}, 10(4):609--631, 2008.

\bibitem{sym2}
Xianjin Chen and Jianxin Zhou.
\newblock A local min-max-orthogonal method for finding multiple solutions to
  noncooperative elliptic systems.
\newblock {\em Math. Comp.}, 79(272):2213--2236, 2010.

\bibitem{sym1}
Xianjin Chen, Jianxin Zhou, and Xudong Yao.
\newblock A numerical method for finding multiple co-existing solutions to
  nonlinear cooperative systems.
\newblock {\em Appl. Numer. Math.}, 58(11):1614--1627, 2008.

\bibitem{mckenna}
Yung~Sze Choi and P.~Joseph McKenna.
\newblock A mountain pass method for the numerical solution of semilinear
  elliptic problems.
\newblock {\em Nonlinear Anal.}, 20(4):417--437, 1993.

\bibitem{terracini2}
M.~Conti, S.~Terracini, and G.~Verzini.
\newblock Nehari's problem and competing species systems.
\newblock {\em Ann. Inst. H. Poincar\'e Anal. Non Lin\'eaire}, 19(6):871--888,
  2002.

\bibitem{conti}
M.~Conti, S.~Terracini, and G.~Verzini.
\newblock An optimal partition problem related to nonlinear eigenvalues.
\newblock {\em J. Funct. Anal.}, 198(1):160--196, 2003.

\bibitem{cdn}
David~G. Costa, Zhonghai Ding, and John~M. Neuberger.
\newblock A numerical investigation of sign-changing solutions to superlinear
  elliptic equations on symmetric domains.
\newblock {\em J. Comput. Appl. Math.}, 131(1-2):299--319, 2001.

\bibitem{dancer}
E.~N. Dancer, Juncheng Wei, and Tobias Weth.
\newblock A priori bounds versus multiple existence of positive solutions for a
  nonlinear {S}chr\"odinger system.
\newblock {\em Ann. Inst. H. Poincar\'e Anal. Non Lin\'eaire}, 27(3):953--969,
  2010.

\bibitem{gnn}
Basilis Gidas, Wei~Ming Ni, and Louis Nirenberg.
\newblock Symmetry and related properties via the maximum principle.
\newblock {\em Comm. Math. Phys.}, 68(3):209--243, 1979.

\bibitem{zhou1}
Youngxin Li and Jianxin Zhou.
\newblock A minimax method for finding multiple critical points and its
  applications to semilinear elliptic {PDE}'s.
\newblock {\em SIAM Sci.Comp.}, 23:840--865, 2001.

\bibitem{zhou2}
Youngxin Li and Jianxin Zhou.
\newblock Convergence results of a local minimax method for finding multiple
  critical points.
\newblock {\em SIAM Sci. Comp.}, 24:865--885, 2002.

\bibitem{lieb}
Elliott~H. Lieb.
\newblock On the lowest eigenvalue of the {L}aplacian for the intersection of
  two domains.
\newblock {\em Invent. Math.}, 74(3):441--448, 1983.

\bibitem{L-BFGS-B}
J.L. Morales and J.~Nocedal.
\newblock Remark on algorithm 778: L-bfgs-b, fortran subroutines for
  large-scale bound constrained optimization.
\newblock {\em ACM Transactions on Mathematical Software (TOMS)}, 38(1),
  November 2011.

\bibitem{Neuberger97}
John~M. Neuberger.
\newblock A numerical method for finding sign-changing solutions of superlinear
  dirichlet problems.
\newblock {\em Nonlinear World}, 4(1):73--83, 1997.

\bibitem{Neuberger-GNGA}
John~M. Neuberger.
\newblock {GNGA}: recent progress and open problems for semilinear elliptic
  pde.
\newblock {\em Contemp. Math.}, 357:201--237, 2004.
\newblock Amer. Math. Soc., Providence, RI.

\bibitem{Neuberger-Newton}
John~M. Neuberger and James~W. Swift.
\newblock Newton's method and morse index for semilinear elliptic pdes.
\newblock {\em Internat. J. Bifur. Chaos Appl. Sci. Engrg.}, 11(3):801--820,
  2001.

\bibitem{Nocedal-Wright06}
Jorge Nocedal and Stephen~J. Wright.
\newblock {\em Numerical Optimization}.
\newblock Springer Series in Operations Research and Financial Engineering.
  Springer, 2nd edition, 2006.

\bibitem{noris}
Benedetta Noris and Gianmaria Verzini.
\newblock A remark on natural constraints in variational methods and an
  application to superlinear schr\"odinger systems.
\newblock {\em J. Differential Equations}, 254(2013) No. 3, 1529--1547.

\bibitem{rabin}
Paul~H. Rabinowitz.
\newblock {\em Minimax methods in critical point theory with applications to
  differential equations}, volume~65 of {\em CBMS Regional Conference Series in
  Mathematics}.
\newblock Published for the Conference Board of the Mathematical Sciences,
  Washington, DC, 1986.

\bibitem{Triangle}
Jonathan~Richard Shewchuk.
\newblock Delaunay refinement algorithms for triangular mesh generation.
\newblock {\em Comput. Geom.}, 22(1-3):21--74, 2002.
\newblock 16th ACM Symposium on Computational Geometry (Hong Kong, 2000).

\bibitem{szulkin}
Andrzej Szulkin and Tobias Weth.
\newblock Ground state solutions for some indefinite variational problems.
\newblock {\em J. Funct. Anal.}, 257(12):3802--3822, 2009.

\bibitem{tt}
N.~Tacheny and C.~Troestler.
\newblock A mountain pass algorithm with projector.
\newblock {\em J. Comput. Appl. Math.}, 236(7):2025--2036, 2012.

\bibitem{terracini}
Hugo Tavares, Susanna Terracini, Gianmaria Verzini, and Tobias Weth.
\newblock Existence and nonexistence of entire solutions for non-cooperative
  cubic elliptic systems.
\newblock {\em Comm. Partial Differential Equations}, 36(11):1988--2010, 2011.

\end{thebibliography}

\end{document}
