/*! This file is auto-generated */ .wp-block-button__link{color:#fff;background-color:#32373c;border-radius:9999px;box-shadow:none;text-decoration:none;padding:calc(.667em + 2px) calc(1.333em + 2px);font-size:1.125em}.wp-block-file__button{background:#32373c;color:#fff;text-decoration:none} Problem 18 Show that the square of a noncen... [FREE SOLUTION] | 91Ó°ÊÓ

91Ó°ÊÓ

Show that the square of a noncentral \(T\) random variable is a noncentral \(F\) random variable.

Short Answer

Expert verified
The square of a noncentral 'T'-distributed random variable is a noncentral 'F'-distributed random variable. This fact comes from the transformation of the 'T' distribution when squared which follows the distribution of 'F'. The noncentrality parameter of the resulting 'F' distribution is the square of that of the original 'T' distribution.

Step by step solution

01

Understanding 'T' and 'F' Distributions

A noncentral 'T' distribution has a noncentrality parameter which measures the degree of noncentrality. An 'F' distribution is a ratio of two independent chi-square random variables divided by their respective degrees of freedom.
02

Squared 'T' Distribution as 'F' Distribution

If \(X\) is a noncentral Student's 't'-distributed random variable with \(v\) degrees of freedom and noncentrality parameter \(δ\) , then \(X^2\) has a noncentral 'F'-distribution with \(1\) and \(v\) degrees of freedom, with noncentrality parameter \(δ^2\). This is because a noncentral 't'-distribution is found as \(X = (Z + δ) / sqrt{V/v}\), where \(Z\) has standard normal distribution, \(V\) has chi-squared distribution with \(v\) degrees of freedom, and \(Z\) and \(V\) are independent. If you square this, you remove the denominator and subsequently the ratio nature of 't'-distribution turning it into an 'F' distribution.
03

Final Connection

By squaring noncentral 'T', we convert a 'T' distributed variable into a noncentral 'F' distributed variable. The noncentrality parameter of the 'F' distribution will be the square of that of the 'T' distribution.

Unlock Step-by-Step Solutions & Ace Your Exams!

  • Full Textbook Solutions

    Get detailed explanations and key concepts

  • Unlimited Al creation

    Al flashcards, explanations, exams and more...

  • Ads-free access

    To over 500 millions flashcards

  • Money-back guarantee

    We refund you if you fail your exam.

Over 30 million students worldwide already upgrade their learning with 91Ó°ÊÓ!

One App. One Place for Learning.

All the tools & learning materials you need for study success - in one app.

Get started for free

Most popular questions from this chapter

Two different teaching procedures were used on two different groups of students. Each group contained 100 students of about the same ability. At the end of the term, an evaluating team assigned a letter grade to each student. The results were tabulated as follows. $$ \begin{array}{rrrrrrr} \text { Group } & \text { A } & \text { B } & \text { C } & \text { D } & \text { F } & \text { Total } \\ \hline 1 & 15 & 25 & 32 & 17 & 11 & 100 \\ 11 & 9 & 18 & 29 & 28 & 16 & 100 \\ \hline \end{array} $$ If we consider these data to be observations from two independent multinomial distributions with \(k=5\), test, at the 5 per cent significance level, the hypothesis that the two distributions are the same (and hence the two teaching procedures are equally effective).

A die was cast \(n=120\) independent times and the following data resulted: $$ \begin{array}{l|cccccc} \text { Spots up } & 1 & 2 & 3 & 4 & 5 & 6 \\ \hline \text { Frequency } & \mathrm{b} & 20 & 20 & 20 & 20 & 40-\mathrm{b} \end{array} $$ If we use a chi-square test, for what values of \(b\) would the hypothesis that the die is unbiased be rejected at the \(0.025\) significance level? (9:

Using the notation of this section, assume that the means satisfy the condition that \(\mu=\mu_{1}+(b-1) d=\mu_{2}-d=\mu_{3}-d=\cdots=\mu_{b}-d\). That is, the last \(b-1\) means are equal but differ from the first mean \(\mu_{1}\), provided that \(d \neq 0\). Let a random sample of size \(a\) be taken from each of the \(b\) independent normal distributions with common unknown variance \(\sigma^{2}\). (a) Show that the maximum likelihood estimators of \(\mu\) and \(d\) are \(\hat{\mu}=X\) and $$d=\left[\sum_{j=2}^{b} X_{\cdot j} /(b-1)-X_{\cdot 1}\right] / b$$ (b) Find \(Q_{6}\) and \(Q_{7}=c d^{2}\) so that when \(d=0, Q_{7} / \sigma^{2}\) is \(\chi^{2}(1)\) and $$\sum_{i=1}^{a} \sum_{j=1}^{b}\left(X_{i j}-X\right)^{2}=Q_{3}+Q_{6}+Q_{7}$$ (c) Argue that the three terms in the right-hand member of part (b), once divided by \(\sigma^{2}\), are stochastically independent random variables with chi-square distributions, provided that \(d=0\). (d) The ratio \(Q_{7} /\left(Q_{3}+Q_{6}\right)\) times what constant has an \(F\) distribution, provided that \(d=0\) ?

Let \(X_{1 j}, X_{2,}, \ldots, X_{a_{j} j}\) represent independent random samples of sizes \(a\), from normal distributions with means \(\mu\), and variances \(\sigma^{2}, j=\) \(1,2, \ldots, b .\) Show that $$ \sum_{j=1}^{b} \sum_{i=1}^{a_{1}}\left(X_{1},-X\right)^{2}=\sum_{j=1}^{b} \sum_{i=1}^{a_{j}}\left(X_{t j}-X_{\cdot j}\right)^{2}+\sum_{j=1}^{b} a_{j}\left(X_{\cdot j}-X\right)^{2} $$ or \(Q^{\prime}=Q_{3}^{\prime}+Q_{4}^{\prime} .\) Here \(X=\sum_{j=1}^{b} \sum_{i=1}^{a_{1}} X_{i j} / \sum_{j=1}^{b} a\), and \(X_{\cdot j}=\sum_{i=1}^{a_{f}} X_{i} / a_{j} .\) If \(\mu_{1}=\mu_{2}=\cdots=\mu_{b}\), show that \(Q^{\prime} / \sigma^{2}\) and \(Q_{3}^{\prime} / \sigma^{2}\) have chi-square distributions. Prove that \(Q_{3}^{\prime}\) and \(Q_{4}^{\prime}\) are stochastically independent, and hence \(Q_{4}^{\prime} / \sigma^{2}\) also has a chi-square distribution. If the likelihood ratio \(\lambda\) is used to test \(H_{0}: \mu_{1}=\mu_{2}=\cdots=\mu_{b}=\mu, \mu\) unspecified and \(\sigma^{2}\) unknown, against all possible alternatives, show that \(\lambda \leq \lambda_{0}\) is equivalent to the computed \(F \geq c\), where $$F=\frac{\left(\sum_{j=1}^{b} a_{j}-b\right) Q_{4}^{\prime}}{(b-1) Q_{3}^{\prime}}$$ What is the distribution of \(F\) when \(H_{0}\) is true?

A certain genetic model suggests that the probabilities of a particular trinomial distribution are, respectively, \(p_{1}=p^{2}, p_{2}=2 p(1-p)\), and \(p_{3}=\) \((1-p)^{2}\), where \(0

See all solutions

Recommended explanations on Math Textbooks

View all explanations

What do you think about this solution?

We value your feedback to improve our textbook solutions.

Study anywhere. Anytime. Across all devices.