I was reading the section of Data Structures for Disjoint Sets from the text Introduction to Algorithms by Cormen et. al. I made it through the proof, but I'm not sure I understand the very last sentence (and it seems a little handwavy to me)
The prerequisites:
$$\alpha(n) : \text{Inverse of a very fast growing function $A_k(j)$}$$
$$level(x)=\max\{k:x.p.rank\geq A_k(x.rank)\}$$ $$iter(x)=\max\{i:x.p.rank\geq A_{level(x)}^{(i)}(x.rank)\}$$ $$\phi_q(x) = \begin{cases} \alpha(n)\cdot x.rank &\quad\text{if $x$ is a root or $x.rank=0$ }\\ (\alpha(n)-level(x))\cdot x.rank-iter(x) &\quad\text{if $x$ is not a root and $x.rank\geq1$ }\\ \end{cases}$$
Corollary 21.5 : As we follow the simple path from any node toward a root, the node ranks strictly increase. ■
Lemma 21.10 : Let $x$ be a node that is not a root, and suppose that the $q$ th operation is either a $\text{Link}$ or $\text{Find-Set}$. Then after the $q$th operation, $\phi_q(х) \leq \phi_{q-1}(х)$. Moreover, if $x.rank \geq 1$ and either $level(x)$ or $iter(x)$ changes due to the $q$ th operation, then $\phi_q(х) < \phi_{q-1}(х) - 1$. That is, $x$'s potential cannot increase, and if it has positive rank and either $level(x)$ or $iter(x)$ changes, then $x$'s potential drops by at least $1$.■
Now in the proof below, I am confused by the very last sentence (the highlighted portion)
Lemma 21.13: The amortized cost of each $\text{Find-Set}$ operation is $O(\alpha(n))$.
Proof: Suppose that the $q$ th operation is a $\text{Find-Set}$ and that the find path contains $s$ nodes. The actual cost of the $\text{Find-Set}$ operation is $O(s)$. We shall show that no node's potential increases due to the $\text{Find-Set}$ and that at least $\max\{0,s - (\alpha(n) + 2)\}$ nodes on the find path have their potential decrease by at least $1$.
To see that no node's potential increases, we first appeal to Lemma 21.10 for all nodes other than the root. If $x$ is the root, then its potential is $\alpha(n) \cdot x.rank$, which does not change.
Now we show that at least $\max\{0,s - (\alpha(n) + 2)\}$ nodes have their potential decrease by at least $1$. Let $x$ be a node on the find path such that $x.rank > 0$ and $x$ is followed somewhere on the find path by another node $у$ that is not a root, where $level(y) = level(x)$ just before the $\text{Find-Set}$ operation. (Node $у$ need not immediately follow $x$ on the find path.) All but at most $\alpha(n) + 2$ nodes on the find path satisfy these constraints on $x$. Those that do not satisfy them are the first node on the find path (if it has rank $0$), the last node on the path (i.e., the root), and the last node $w$ on the path for which $level(w) = k$, for each $k = 0,1,2,..., \alpha(n) - 1$.
Let us fix such a node $x$, and we shall show that $x$'s potential decreases by at least $1$. Let $k = level(x) = level(y)$. Just prior to the path compression caused by the $\text{Find-Set}$, we have
$x.p.rank \geq A_k^{(iter(x))}(x.rank)$ (by definition of $iter(x)$),
$y.p.rank \geq A_k(y.rank)$ (by definition of $level(y)$),
$y.rank \geq x.p.rank$ (by Corollary 21.5 and because $у$ follows $x$ on the find path)
Putting these inequalities together and letting $i$ be the value of $iter(x)$ before path compression, we have
$y.p.rank \geq A_k(y.rank) \geq A_k(x.p.rank)$ (because $A_k(j)$ is strictly increasing) $ \geq A_k(A_k^{(iter(x))}(x.rank)) = A_k^{(i+1)}(x.rank)$ .
Because path compression will make $x$ and $у$ have the same parent, we know that after path compression, $x.p.rank = y.p.rank$ and that the path compression does not decrease $y.p.rank$. Since $x.rank$ does not change, after path compression we have that $x.p.rank\geq A_k^{(i+1)}(x.rank)$. Thus, path compression will cause either $iter(x)$ to increase (to at least $i + 1$) or $level(x)$ to increase (which occurs if $iter(x)$ increases to at least $x.rank + 1$). In either case, by Lemma 21.10, we have $\phi_q(х) \leq \phi_{q-1}(х) - 1$. Hence, $x$'s potential decreases by at least $1$.
The amortized cost of the $\text{Find-Set}$ operation is the actual cost plus the change in potential. The actual cost is $O(s)$, and we have shown that the total potential decreases by at least $\max\{0,s - (\alpha(n) + 2)\}$. The amortized cost, therefore, is at most $O(s) - (s - (\alpha(n) + 2)) = O(s) - s + O(\alpha(n)) = O(\alpha(n))$, $\require{color}\colorbox{yellow}{since we can scale up the units of potential to dominate the constant hidden in $О(s)$.}$ ■
The authors elaborate later: The proof of Lemma 21.13 ends with scaling the units of potential to dominate the constant hidden in the term. To be more precise in the proof, you need to change the definition of the potential function to multiply each of the two cases by a constant, say $c$, that dominates the constant in the term $O(s)$
Can anyone help me understand what the authors mean?
The corresponding portion of the text can be found here