3

Of course, Gaussian elimination is safe to use as proven by the countless systems I've solved with it while practising my linear algebra (which is I must add very basic and low-level), but when Jim Hefferon's free textbook on linear algebra raised the question why

  1. Scaling rows by a non-zero constant
  2. Adding rows to each other

Do not change the solution set, I found myself incapable of giving correct mathematical proof.

In his answer manual, Hefferon himself "proves" the safety of both operations by showing that each operation can be reversed without adding or losing solutions.

For example, if a row is scaled by a nonzero constant C, this operation can be reversed by dividing both sides by C, without losing or creating solutions.

Does this satisfy as mathematical proof? To me it seems like proving and operation through its usage isn't exactly proving its correctness, because it makes the assumption that the solution set wasnt changed between performing and reversing the operation. If this indeed does not satisfy as proof, then what would be proof that no solutions are lost performing Gaussian elimination?

  • 1
    What kind of proof are you looking for then? It's pretty clear you can do whatever you want to linear equations so long as you don't introduce extraneous solutions or change one side without changing the other. Same goes for adding one row to another: so long as you add each side to their corresponding sides for whatever row you're adding to, then everything should work out just fine. Are you looking for a non-algebraic proof of some sort? – Daniel W. Farlow Aug 12 '15 at 14:54
  • I would prove that 1. If $s$ is a solution of the original equation $f(x)=0$, then it is still a solution of $Cf(x)=0$; 2. If $t$ is not a solution of the original solution, then it is still not a solution of $Cf(x)=0$. – KittyL Aug 12 '15 at 14:56
  • If you multiply one equation by a non-zero constant, then any solution of the original set is still a solution. But then you can multiply the new equation by the reciprocal of that constant in order to conclude that a solution of the second set of equations must be a solution of the first set. Next, if you add to an equation $Eq_a$ another $Eq_b$ to obtain a set where $E_a$ is replaced by $Eq_a+Eq_b$, solutions of the first set remain solutions of the second. And, solutions of the second are solutions of the first because $(Eq_a+E_b)-Eq_b$ gets you back to $Eq_a$. – Disintegrating By Parts Aug 12 '15 at 16:53

3 Answers3

3

ALmost. Maybe the best way to see this is as follows: If $C$ is an invertible matrix that the set $S_1$ of vectors $x$ such that $Ax=b$ is the same as the set $S_2$ of vectors $x$ such that $CAx=Cb$. This follows because for each $x$ with $Ax=b$ we immediately find $C\cdot Ax=Cb$ and vice verse for each $x$ with $CAx=Cb$ we find by multiplying with $C^{-1}$ (which exists by the assumption of invertibility) $Ax=C^{-1}CAx=C^{-1}Cb=b$. Therefore we have bothe $S_1\subseteq S_2$ and $S_2\subseteq S_1$.

Now to apply trhis to Gauss elimination observe that the single steps (scaling a row, adding a row to another row, swapping rows) can be accieved by multiplying with a suitable simple matrix $C$ with an easily found inverse.

1

It is a little hard to write all the equations down but I'll try to explain the process.

There are three fundamental operations we perform on a linear system.

  • Multiplying a row by a scalar.
  • Interchanging two rows.
  • Adding a scalar multiple of a row to another a row.

The thing to notice is that after performing any of these three operations the resultant system consists of equations that are linear combinations of the original one.

For an example say we have a system,

\begin{array}{c} a_1x+b_1y+c_1z=d_1 \\ a_2x+b_2y+c_2z=d_2 \\ a_3x+b_3y+c_3z=d_3 \end{array}

Say we multiply the first row by $q$ and add it to the second. The resultant system is,

\begin{array}{c} a_1x+b_1y+c_1z=d_1 \\ q(a_1x+b_1y+c_1z) + a_2x+b_2y+c_2z=q \cdot d_1 + d_2 \\ a_3x+b_3y+c_3z=d_3 \end{array}

Now say $(x, y, z)^T$ was a solution to the original system. Then, it is also a solution to the second one. Let us try to convince ourselves of this. The first and last rows are not a problem. The second equation of the resultant system is also satisfied because $a_2x+b_2y+c_2z= d_2$ and $(a_1x+b_1y+c_1z) = d_1 \implies q (a_1x+b_1y+c_1z) = q \times d_1 $.

The other two linear operations are also similarly disposed of.

Now look at what we have proven. We have proven that "any solution to the original system is a solution to the system resulting from one of the three linear operations".

But we require a little more. We want the solutions of the new system to be exactly those of the first one. This is established by the fact that every linear operations mentioned has a corresponding inverse operation which is also one of the three linear operations. For an example the inverse operation of the one we performed above is multiplying the first row of the second system by $-q$ and adding to the second row. Now think of the original system as resulting from the second one through the performance of a linear operation. Hence from what we proved above any solution to the second system is also a solution to the first.

So any solution to the original system is a solution to the resultant system and any solution to the resultant system is a solution to the original one. Hence the solutions to the first system are exactly those to the second.

This is exactly what we require.

There is a nice explanation of all this in the first twenty or so pages in "Linear Algebra by Hoffman, Kunze". Must read.

Ishfaaq
  • 10,230
  • While I thank you for the proof of the first statement, I still have some doubts about the second statement, though. In the end, you prove that a third system resulting from the second system has the same set of solutions as the first system, thus assume all three systems have the same solutions. But why couldn't one argue that the second system could have picked up some extra solutions through its operation and then lost them through the inverse operation? It seems to me you make the assumption that as the solution to the first and third systems are equal, the second one must share it too. – Heatherfield Aug 12 '15 at 17:01
  • I'm not trying to disprove Gauss, neither am I trying to act like a marhematical conspiracy theorist, but to me it seems like the safety of the operations COULD be proven, or at least dumbed down to rules from singlevariable algebra. – Heatherfield Aug 12 '15 at 17:15
  • What I have said is that you can attain the "first" system (not a third different one) from the second through a linear operation too. Take another read mate. I think it's clear enough. There is an answer to exactly what you ask in each of the responses. – Ishfaaq Aug 12 '15 at 17:17
  • I failed to link the second proof to the first proof mentally, mistake on my end. Thanks for the clear response – Heatherfield Aug 12 '15 at 17:48
0

Instead of thinking about the linear algebra (i.e., matrix/vector) justification, I like to look at the underlying algebra problem (i.e., the system of linear equations). In the world of the algebraic equations, the Gaussian Elimination (GE) on the linear algebra structures corresponds to the rules you learn when first trying to solve an equation. Namely,

  1. adding the same quantity to both sides of the equation does not change the solution.
  2. multiplying both sides of the equation by a constant different from zero does not change the solution.

If you take these rules for granted, then you must believe GE is simply a compact way to write down the operations that you do on the system of linear equations, using the language of matrices. In particular, adding two rows is fine, since the quantities you add on the left hand side and right hand side are equal (since they satisfy another one of the equations of the system).

Matrices/vectors are an interesting world. However, when it comes to their use in linear systems, I like to think of them simply as a nice compact way to write down something that otherwise would take a lot of space on the paper, allowing to have a good vision of the big picture, without getting lost in the details of each single component.

bartgol
  • 6,361