This topic has been discussed in this forum, but I don't think the problem has been addressed completely.
To apply Pearson's test to a contingency table one computes $$\sum{ {\rm (Observed - Expected)}^2 \over {\rm Expected}}$$ and argues that under certain conditions this statistic has an approximate $\chi^2$ distribution.
The obvious question is: why don't we divide by Expected$^2$? The answer that in this way we get an (approximate) sum of standard normals in my view is not acceptable: we should do what must be done and if what we get is not nice, well, so be it. In fact, what can happen is this. Suppose that we take the original sample and split it in two, keeping the same proportions, i.e. the count of each cell is the same proportion of the total count as it was before splitting. In this case, the value of the statistic is half of what it was before, and it may well happen that a non-rejected H null is now rejected. How is this reasonable? Shouldn't the dependence / independence determination be the same in both cases? I haven't found a discussion of this aspect anywhere in the literature. Pointers would be welcome. Thanks.