Questions tagged [chi-squared]

Use this tag for questions about (1) distributions of a sum of squares of independent standard normal random variables or (2) statistical hypothesis tests with such a sampling distribution if the null hypothesis is true.

Distribution

In probability theory and statistics, the chi-squared distribution (also chi-square or χ$^2$ distribution) with k degrees of freedom is the distribution of a sum of the squares of k independent standard normal random variables. The chi-squared distribution is a special case of the gamma distribution and is one of the most widely used probability distributions in inferential statistics, notably in hypothesis testing or in construction of confidence intervals.

Test

A chi-squared test is any statistical hypothesis test for which the sampling distribution of the test statistic is a chi-squared distribution if the null hypothesis is true. The chi-squared test is used to determine whether there is a significant difference between the expected frequencies and the observed frequencies in one or more categories.

In standard applications of the test, observations are classified into mutually exclusive classes, and there is a null hypothesis that gives the probability that any observation falls into the corresponding class. The purpose of the test is to evaluate how likely the observations would be assuming the null hypothesis is true.

Chi-squared tests are often constructed from a sum of squared errors or through the sample variance. Test statistics that follow a chi-squared distribution arise from an assumption of independent normally-distributed data, which is valid in many cases. A chi-squared test can be used to attempt rejection of the null hypothesis that the data are independent.

441 questions
16
votes
3 answers

Proof and precise formulation of Welch-Satterthwaite equation

In my statistics course notes the Welch-Satterthwaite equation, as used in the derivation of the Welch test, is formulated as follows: Suppose $S_1^2, \ldots, S_n^2$ are sample variances of $n$ samples, where the $k$-th sample is a sample of size…
9
votes
2 answers

Why the chi-squared statistic follows chi-squared distribution?

The formula for the Chi-Square test statistic is the following: $$\chi^2=\sum_{i=1}^n\frac{({O_i-E_i})^2}{E_i}$$ where $O_i$ is observed data, and $E_i$ is expected. I am just curious why this follows the $\chi^2$ distribution?
8
votes
1 answer

Chi-square goodness of fit test proof

I understand the classcial $\chi^2$ "goodness of fit" test used in Statistics, in which we compute $\sum_{i=1}^n \frac{(O_i - E_i)^2}{E_i}$ and, by comparing this quantity to a value found in a table of $\chi^2$ law (with a given risk $\alpha = 5\%$…
Basj
  • 1,591
7
votes
1 answer

Chi Squared for Goodness of Fit

Hi, any help is appreciated :) I am trying to teach myself statistics. I've watched the Khan Academy Series on Chi square statistic for hypothesis testing…
7
votes
0 answers

Confusion regarding usage of Mahalanobis distance for update rejection in Kalman filtering

I recently came across some material that discussed a method for performing update rejection in Kalman filters when bad measurements are received. [Paper 1] [Paper 2: see Section III(E)] This method involves the usage of Mahalanobis distance to…
7
votes
1 answer

Variance of variance MLE estimator of a normal distribution

The MLE estimator of the variance of a normal distribution is $\hat \sigma^2 = \frac{1}{N} \sum_{i=1}^{N}(X_i - \bar X)^2$, where $\bar X$ is the sample mean and $X_i \sim^{iid} \mathcal{N}(\mu,\sigma^2)$ . I'm curious because I've seen (e.g.…
6
votes
1 answer

Why does the Chi-squared test statistic follow the Chi-squared distribution

I know the Chi-squared test statistic is defined as: $$\chi^2=\sum_{i=1}^n\frac{({O_i-E_i})^2}{E_i}$$ where $O_i$ is observed data, and $E_i$ is expected. I also know that the $\chi^2$ distribution is essentially defined as the sum of squared…
6
votes
1 answer

Deriving the Chi-squared distribution using characteristic functions

I would like to directly derive the probability density function (PDF) for a Chi-squared distribution with $k$ degrees of freedom using characteristic functions. If $X_{1}, X_{2}, \dots, X_{k}$ are independent, standard normal random variables,…
6
votes
2 answers

Chi Square Contingency Table - Formula Derivation

A chi-square distribution is constructed from normal random variables $X_{i=1, \dots ,n}$, each with normal distribution and mean $\mu$ and variance $\sigma^2$. Transforming to standard normal and squaring, i.e.: $$\frac{(X_i -…
6
votes
0 answers

Variance of the Euclidean norm of a vector of Gaussians

Given a vector of correlated Gaussian random variables $\vec{X}=\left(X_1, ..., X_k\right)$ that are normally distributed $\vec{X} \sim \mathcal{N}\left(\vec{\mu}, \Sigma\right)$ with arbitrary $\Sigma$ and $\vec{\mu}$ values what's the variance…
5
votes
1 answer

Inverse Gaussian distribution

Let $X_1,....X_n$ be a random sample from the inverse Gaussian distribution with pdf $$f(x|\mu,\lambda)=\left(\frac{\lambda}{2\pi x^3}\right)^{1/2} \exp\left(\frac{-\lambda (x-\mu)^2}{2 \mu^2 x}\right),\quad0 \lt x \lt \infty.$$ Let…
Yang
  • 531
  • 1
  • 9
  • 18
5
votes
1 answer

Is the probability of $X$ in the interval $[0, \mathbb{E}{[X]}]$ at least $1/2$, if $X \sim \chi^2$?

I am working on the $\chi^2$ distribution and have the following assumption: The cumulative distribution function of a $\chi^2$ distributed random variable is greater than $\frac{1}{2}$ at the right boundary of the interval $[0,…
5
votes
1 answer

Bound the probability of multivariate gaussian vector norm.

Let's say $v \in \mathbb{R}^n \sim \mathcal{N}(0, \sigma I)$. That is, $v$ is a gaussian random vector, whose entries are distributed $\mathcal{N}(0, \sigma)$ i.i.d. From the book "C. Giraud. Introduction to high-dimensional statistics", it can be…
5
votes
0 answers

Why should the sum of squares of two Independent normals be memory-less

In section 11.3.1 of Introduction to probability models by Ross (10th edition), a very strange phenomenon is described. If you take two independent standard normal distributions and sum their squares, you get an exponential distribution with rate…
5
votes
3 answers

Why can we not accept the alternate hypothesis in Chi Squared Testing?

I'm a math teacher, but this aspect of stats is not my strong point. I've asked several other teachers as to why, and their responses was just "don't do it" the why was not very compelling, so I come here. $H_\text{null}$ = $m$ and $n$ are…
1
2 3
29 30