I have the following problem where I have already established convergence in the general case. I would like to show the convergence in a lower-dimensional space spanned by basis functions.
General problem
Consider the following recursion and establish that the algorithm will converge to a unique solution $ v^* \in \mathbb{R}^{n}$ from any initial guess $v_0$ and small enough time step $\Delta t$ as $ t \to + \infty$:
$$ v_{t+\Delta t} \left ( \mathbf{x} \right ) = G(v_{t} \left ( \mathbf{x} \right )), \quad \mathbf{x} \in \Omega \subset \mathbb{R}^{n \times k} $$
where $\Omega$ is a compact set, $v_{t} \left ( \mathbf{x} \right ) = \left [ v \left ( x^{(1)} \right ) , \cdots , v \left ( x^{(n)} \right ) \right ]' $ is the $n$-dimensional vector of images of each matrix lines $x^{(i)} \in \mathbb{R}^{k}$ by $v_{t}$ and G is a continuously-differentiable non-linear map $\mathbb{R}^n \mapsto \mathbb{R}^n$.
For more context, the $x^{(i)}$'s are the $n$ points used for the discrization of a $k$-dimensional state-space and, $v$ is the value function that I need to solve, and $G$ results from the discretization of some PDE $ F \left ( \mathbf{x} , \mathbf{v} , \triangledown \mathbf{v} , \triangledown^2 \mathbf{v} \right )$ using a FD scheme. I want to solve for the solution to an HJB equation using a time-marching algorithm, which is a fixed-point of the above equation. You can abstract from the $\mathbf{x}$'s for now (I come back to them later):
$$ \mathbf{\mathbf{v}}_{t+\Delta t} = G \left ( \mathbf{v}_{t} \right ), \quad \mathbf{v}_t \in \mathbb{R}^{n} $$
Proof: In my setup, I can show that the Jacobian is of the following form
$$ \triangledown_v G(\mathbf{v}) = \left [ (1+\rho \Delta t )\mathbb{I}_n - A \Delta t \right ]^{-1} $$
where $\rho>0$, $A$ is an $n \times n$ matrix, and it is zero column-sum and has only non-negative off-diagonal elements, i.e.
$$ \sum_{j=1}^{N} a_{ij} = 0 , \forall i \quad \text{ and } \quad \left| a_{ij} \right| = a_{ij} , \forall i \neq j $$
Let $B = (1+\rho \Delta t )\mathbb{I}_n - A \Delta t $. By Gershgorin circle theorem, one can show that all eigenvalues $\lambda_i$ of $B$ are bounded below by $1+\rho \Delta t$ as they are contained in the circles of center $1+\rho \Delta t - \Delta t a_{ii}$ and of radius $R_i = \Delta t |a_{ii} | $. Therefore, $ \mathbb{I}_n - A \frac{\Delta t}{1+\rho \Delta t} $ is invertible and
$$ \left\| \triangledown_v G(\mathbf{v}) \right\|_2 \simeq \left\| \left ( (1+\rho \Delta t )\mathbb{I}_n - A \Delta t \right )^{-1} \right\| _2 + o \left ( \Delta t^2 \right ) < 1 + o \left ( \Delta t^2 \right ) $$
Hence the algorithm converges quadratically to the solution $\mathbf{v}^*$ provided that $\Delta t$ is small enough.
Modified problem
Typically in the above problem, $n$ is very large because it corresponds to the grid points used for the discretization of $\Omega_{dense}$.
I now want to reduce the dimensionality of this problem by approximating the values $\left \{ v_{i} \right \}_1^n$ using $M$ basis functions $\left\{ \phi_1(.), \cdots , \phi_m(.) \right\}$ mapping $\Omega_{dense}$ into $\mathbb{R}$ (I chose local polynomials in my case). Typically, $M << n$. Let $\tilde{V}$ be the space spanned by my basis functions.
$$ \tilde{V} = \left\{ \tilde{v} \in \mathbb{R} : \tilde{v} = \sum_{m=1}^{M} \phi_m\left ( \mathbf{x} \right ) w_m = \Phi \left ( \mathbf{x} \right ) \cdot \mathbf{w}, \mathbf{w} \in \mathbb{R}^m \text{ and } \mathbf{x} \in \Omega_{dense} \right\} $$
Now, the problem rewrites
$$ \widetilde{\mathbf{v}}_{t+\Delta t} = G(\tilde{\mathbf{v}}_t), \quad \widetilde{\mathbf{v}}_t \in \widetilde{V} $$
where G is the same non-linear map. This problem can be reduced to finding a fixed-point $\mathbf{w}^*$ of the $\mathbf{w}_t$ coefficients in $\mathbb{R}^M$:
$$ \mathbf{w}_{t+\Delta t} = \tilde{G}(\mathbf{w}_t), \quad \mathbf{w}_t \in \mathbb{R}^m $$
I want to use the same dataset of $\left \{ v(x_i), 1 \leq i \leq n \right \}$ therefore the above problem becomes overdetermined. I again can show that the Jacobian has a convenient expression:
$$ \triangledown_w \tilde{G} \left ( \mathbf{w} \right ) = \tilde{B} ^{\dagger} \Phi $$
where $\tilde{B} = B \Phi = \left [ (1+\rho \Delta t )\mathbb{I}_n - A \Delta t \right ] \Phi$, $\tilde{B} ^{\dagger} = \left [ \tilde{B}' \tilde{B} \right ]^{-1}\tilde{B}'$ is the pseudo-inverse $\Phi = \left [ \phi_m \left ( x_n \right ) \right ]_{n,m} \in \mathbb{R}^{n \times M}$ is rectangular. I effectively run a least-square regression at every iteration and project the $n$ values of $v$ onto the space of real coefficients with dimension $M$.
Question
I would like to establish that this modified system still converges to a unique solution $w^*$ but despite different attempts, I get stuck on the fact that $\Phi$ is rectangular and complicates the problem.
Would you have any ideas on how to prove this?
My attempts led me to noticing the following
- $A$ still is zero column-sum $B$ still has all its eigenvalues bounded below.
- $\tilde{B} = B \Phi$ is full column-rank $M$
- $\tilde{B}$ admits a singular value decomposition that should have singular values closely related to that of B (my guess). See here
- If the basis $\phi^m (.)$ are orthonormal, one should have that $\phi_m (x) \phi_{m'} (x) = 0 , \forall m \neq m', \forall x$ and $\sum_{m=1}^{M} \phi_m(x) = 1 , \forall x$ (I also didn't prove this). For now, my basis functions are not even orthogonal but I wondered if one could use a change of basis argument.
- $A'A$ is both zero column- and row-sum so its eigenvalues are all $0$