I want to have an intuition for why AB in matrix multiplication is not same as BA. It's clear from definition that they are not and there are arguments here (Fast(est) and intuitive ways to look at matrix multiplication?) that explain that this is in order to maintain certain compositional properties. Example:
(column vector)
$A = \left( \begin{array}{c} 1\\ 2\\ 3 \end{array} \right)$
(row vector)
$B = \left(1, 5, 0\right)$
If we view a matrix as a linear combination of columns, then I read this from right to left as saying "take 1 times" of column 1 of left matrix, then "take 5 times" of column 2 of left matrix, then "take 0 times" of column 3 of left matrix. Intuitively this means the $B$ vector is the set of weights for the linear combination and the columns of $A$ are the ones being combined. This yields:
$AB = \left( \begin{array}{c} 1 & 5 & 0\\ 2 & 10 & 0\\ 3 & 15 & 0 \end{array} \right)$
1st question: is this a valid way to think of the operation? It gives the right answer here, but more generally is it correct?
2nd question: how can we apply this (or a better) intuition to the case of mutiplying $BA$? We have:
$BA = \left((1\times1) + (5\times2) + (3\times0)\right) = \left(11\right)$
not sure how to think of that intuitively.
one intuition that has been proposed is matrix multiplication as linear composition of functions. I'm open to that but usually I don't think of matrices like $A$ and $B$ as individually representing functions.