Edit: I've decided to flesh this out to motivate Furstenberg's theorem. This is not the ideal solution, but it is a classical answer to the problem (see https://link.springer.com/article/10.1007/BF00537227)
Consider the space $X := \{1,2\}^\mathbb{Z}$ and let $\sigma :X \rightarrow X$ be the usual left shift map. Define
$$ A : X \rightarrow \text{SL}(2,\mathbb{R}), \ \ A(x) := \begin{cases} A_1 &\text{ if } x_0 = 1 \\ A_2&\text{if } x_0 = 2. \end{cases} $$
This generates a cocycle as follows:
$$ A : X \times \mathbb{Z} \rightarrow \text{SL}(2,\mathbb{R}), \ \ A(x,n) := A(\sigma^{n-1}(x)) \cdots A(x) \text{ if } n > 0,$$
and you take the appropriate definition for $n < 0$. A constant $0 < p < 1$ generates a probability measure $\mu$ on $X$ in the usual sense; for example, if I fix coordinates of $x$ with $n$ $1$'s and $m$ $2$'s, then the probability of such a sequence occurring is $p^n (1-p)^m$.
Define
$$ \lambda : X \times \mathbb{R}^2 \rightarrow \mathbb{R}, \ \ \lambda(x,v) := \limsup_{n \rightarrow \infty} \frac{1}{n}\log(\|A(x,n)v\|).$$
Oseledets' theorem tells us that if $\lambda > 0$ then there is a distribution $x \mapsto E^-_x$ so that for almost every $x \in X$ the following holds:
$$\lambda(x,v) = \begin{cases} \lambda &\text{if } v \in \mathbb{R}^2 \setminus E^-_x\\ - \lambda &\text{if } v \in E^-_x \setminus \{0\}.\end{cases}$$
This distribution is nice, in the sense that it is invariant and measurable. Since every non-zero point in this distribution gives the same value, this suggests that instead of thinking about the vector space, we should maybe think about projective space (or the space of directions). For $v,w \in \mathbb{R}^2 \setminus \{0\}$, we write $v \sim w$ if $v = \lambda w$ with $\lambda \in \mathbb{R}$. Define projective space by
$$ \mathbb{P}^2 := (\mathbb{R}^2 \setminus \{0\})/\sim.$$
Abusing notation, it is clear that our matrices act on projective space by $A_i([v]) := [A_i(v)]$, where here $[v] := \{w \in \mathbb{R}^2 \ | \ v \sim w\}.$ Now, if we abuse Oseledets' theorem, we can see that our Lyapunov exponent $\lambda$ is really a function on $X \times \mathbb{P}^2$, and instead of a distribution we're really thinking about a "nice" function $p_- : X \rightarrow \mathbb{P}^2$ which determines a direction that corresponds to $-\lambda$ (and to make it worse, all of this is almost everywhere). To be precise, we now have
$$ \lambda : X \times \mathbb{P}^2 \rightarrow \mathbb{R}, \ \ \lambda(x,p) := \limsup_{n \rightarrow \infty} \frac{1}{n}\log \left(\|A(x,n)v\| \right) \text{ for some } v \in p.$$
It is an easy exercise to show it is well-defined, independent of the choice of $v \in p$.
Let $p_1, p_2 \in \mathbb{P}^2$ correspond to the eigenspaces of $A_1$. Notice that $\{p_1, p_2\}$ is the only invariant set for the cocycle $A$, so if we are going to have our distribution $p_-$ from earlier, then it must live in here. Notice also that we have the relations
$$A_1(p_1) = p_1, \ A_1 (p_2) = p_2, \ A_2(p_2) = p_1, \ A_2 (p_1) = p_2.$$
Let $C_i := \{x \in X \ | \ x_0 = i\}$, so $X = C_1 \sqcup C_2$.
Now, if $x \in C_1$, then
$$ \lambda(x,p_1) = \limsup_{n \rightarrow \infty} \frac{1}{n}\log \left( \|A(\sigma(x), n-1) A_1 v\|\right) \text{ for some } v \in p_1.$$
Say $p_1$ is the eigenspace corresponding to $\sigma$, so $A_1 v = \sigma v$. This gives us
$$ \lambda(x,p_1) = \limsup_{n \rightarrow \infty} \frac{1}{n}\log \left( \sigma \|A(\sigma(x), n-1) v\| \right) \text{ for some } v \in p_1.$$
Simplifying this yields $\lambda(x, p_1) = \lambda(\sigma(x), p_1).$ A similar argument yields that if $x \in C_2$, then $\lambda(x, p_1) = \lambda(\sigma(x), p_2)$. Using a conditioning trick (i.e. noting $\lambda$ relies only on future coordinates), we have
$$ \int_X \lambda(x,p_1) d\mu(x) = \int_{C_1} \lambda(x, p_1) d\mu(x) + \int_{C_2} \lambda(x, p_1) d\mu(x) \\ = \int_{C_1} \lambda(\sigma(x), p_1) d\mu(x) + \int_{C_2} \lambda(\sigma(x), p_2) d\mu(x) \\ = p\int_{X} \lambda(x,p_1) d\mu(x) + (1-p)\int_{X} \lambda(x, p_2) d\mu(x).$$
We conclude that
$$\int_X \lambda(x,p_2) d\mu(x) = \int_X \lambda(x,p_1) d\mu(x)$$
as long as $0 < p < 1.$ Supposing now that $\lambda > 0$, we have that there is a set of full measure so that $\lambda(x, p_-(x)) = - \lambda$. On the other hand, we noted earlier that $p_-(x) \in \{p_1, p_2\}$. Let $E_i = \{x \in X \ | \ p_-(x) = p_i\}$, and note that $E_1 \sqcup E_2 = X$. We have
$$ (1- 2\mu(E_1)) \lambda = \int_X \lambda(x, p_1) d\mu(x) = \int_X \lambda(x, p_2) d\mu(x) = (1 - 2\mu(E_2))\lambda.$$
Rearrange to get
$$ \mu(E_1) = 1/2 = \mu(E_2)$$
as long as $\lambda \neq 0$. So, half the time $p_-(x)$ is equal to $p_1$, and the other half it is equal to $p_2$. But notice that this implies
$$ \int_X \lambda(x,p_1) d\mu(x) = \int_{E_1} \lambda(x,p_1) d\mu(x) + \int_{E_2} \lambda(x, p_2) d\mu(x) = 0.$$
On the other hand,
$$ - \lambda = \int_X \lambda(x, p_-(x)) d\mu(x) = \frac{1}{2}\int_{X} \lambda(x, p_1) d\mu(x) + \frac{1}{2}\int_X \lambda(x, p_2)d\mu(x) = 0.$$
Notice that in some sense this recreates the stationary measure calculation.