<< Chapter < Page | Chapter >> Page > |
Suppose you measure a collection of scalars ${x}_{1},,{x}_{N}$ . You believe the data is distributed in one of two ways. Your first model, call it ${H}_{0}$ , postulates the data to be governed by the density ${f}_{0}(x)$ (some fixed density). Your second model, ${H}_{1}$ , postulates a different density ${f}_{1}(x)$ . These models, termed hypotheses , are denoted as follows: $${H}_{0}:({x}_{n}, {f}_{0}(x)),n=1N$$ $${H}_{1}:({x}_{n}, {f}_{1}(x)),n=1N$$ A hypothesis test is a rule that, given a measurement $x$ , makes a decision as to which hypothesis best "explains" the data.
Suppose you are confident that your data is normally distributed with variance 1, but you are uncertain aboutthe sign of the mean. You might postulate $${H}_{0}:({x}_{n}, (-1, 1))$$ $${H}_{1}:({x}_{n}, (1, 1))$$ These densities are depicted in .
Assuming each hypothesis isThe concepts introduced above can be extended inseveral ways. In what follows we provide more rigorous definitions, describe different kinds of hypothesis testing, andintroduce terminology.
In the most general setup, the observation is a collection ${x}_{1},,{x}_{N}$ of random vectors. A common assumption, which facilitates analysis, is that the data are independent and identicallydistributed (IID). The random vectors may be continuous, discrete, or in some cases mixed. It is generally assumedthat all of the data is available at once, although for some applications, such as Sequential Hypothesis Testing , the data is a never ending stream.
When there are two competing hypotheses, we refer to a binary hypothesis test. When the number of hypotheses is $M\ge 2$ , we refer to an M-ary hypothesis test. Clearly, binary is a special case of $M$ -ary, but binary tests are accorded a special status for certain reasons. These includetheir simplicity, their prevalence in applications, and theoretical results that do not carry over to the $M$ -ary case.
Suppose we wish to transmit a binary string of length $r$ over a noisy communication channel. We assign each of the $M=2^{r}$ possible bit sequences to a signal ${s}^{k}$ , $k=\{1, , M\}$ where $${s}_{n}^{k}=\cos (2\pi {f}_{0}n+\frac{2\pi (k-1)}{M})$$ This symboling scheme is known as phase-shift keying (PSK). After transmitting a signal across the noisy channel, the receiver faces an $M$ -ary hypothesis testing problem: $${H}_{0}:x={s}^{1}+w$$ $$$$ $${H}_{M}:x={s}^{M}+w$$ where $(w, (0, ^{2}I))$ .
In many binary hypothesis tests, one hypothesis represents the absence of a ceratinfeature. In such cases, the hypothesis is usually labelled ${H}_{0}$ and called the null hypothesis. The other hypothesis is labelled ${H}_{1}$ and called the alternative hypothesis.
Consider the problem of detecting a known signal $s=\left(\begin{array}{c}{s}_{1}\\ \\ {s}_{N}\end{array}\right)$ in additive white Gaussian noise (AWGN). This scenario is common in sonar and radar systems. Denotingthe data as $x=\left(\begin{array}{c}{x}_{1}\\ \\ {x}_{N}\end{array}\right)$ , our hypothesis testing problem is $${H}_{0}:x=w$$ $${H}_{1}:x=s+w$$ where $(w, (0, ^{2}I))$ . ${H}_{0}$ is the null hypothesis, corresponding to the absence of a signal.
Notification Switch
Would you like to follow the 'Statistical signal processing' conversation and receive update notifications?