Chapter 8: Problem 5
If \(X_{1}, X_{2}, \ldots, X_{n}\) is a random sample from a distribution having
pdf of the form \(f(x ; \theta)=\theta x^{\theta-1}, 0
/*! This file is auto-generated */ .wp-block-button__link{color:#fff;background-color:#32373c;border-radius:9999px;box-shadow:none;text-decoration:none;padding:calc(.667em + 2px) calc(1.333em + 2px);font-size:1.125em}.wp-block-file__button{background:#32373c;color:#fff;text-decoration:none}
Learning Materials
Features
Discover
Chapter 8: Problem 5
If \(X_{1}, X_{2}, \ldots, X_{n}\) is a random sample from a distribution having
pdf of the form \(f(x ; \theta)=\theta x^{\theta-1}, 0
All the tools & learning materials you need for study success - in one app.
Get started for free
Illustrative Example \(8.2 .1\) of this section dealt with a random sample of size \(n=2\) from a gamma distribution with \(\alpha=1, \beta=\theta .\) Thus the mgf of the distribution is \((1-\theta t)^{-1}, t<1 / \theta, \theta \geq 2 .\) Let \(Z=X_{1}+X_{2}\). Show that \(Z\) has a gamma distribution with \(\alpha=2, \beta=\theta .\) Express the power function \(\gamma(\theta)\) of Example \(8.2 .1\) in terms of a single integral. Generalize this for a random sample of size \(n .\)
Consider a distribution having a pmf of the form \(f(x ; \theta)=\theta^{x}(1-\theta)^{1-x}, x=\) 0,1, zero elsewhere. Let \(H_{0}: \theta=\frac{1}{20}\) and \(H_{1}: \theta>\frac{1}{20} .\) Use the central limit theorcu? to determine the sample size \(n\) of a random sample so that a uniformly most powerful test of \(H_{0}\) against \(H_{1}\) has a power function \(\gamma(\theta)\), with approximately \(\gamma\left(\frac{1}{20}\right)=0.05\) and \(\gamma\left(\frac{1}{10}\right)=0.90\).
Consider a random sample \(X_{1}, X_{2}, \ldots, X_{n}\) from a distribution
with pdf \(f(x ; \theta)=\theta(1-x)^{\theta-1}, 0
Let the independent random variables \(Y\) and \(Z\) be \(N\left(\mu_{1}, 1\right)\) and \(N\left(\mu_{2}, 1\right)\), respectively. Let \(\theta=\mu_{1}-\mu_{2}\). Let us observe independent observations from each distribution, say \(Y_{1}, Y_{2}, \ldots\) and \(Z_{1}, Z_{2}, \ldots\) To test sequentially the hypothesis \(H_{0}: \theta=0\) against \(H_{1}: \theta=\frac{1}{2}\), use the sequence \(X_{i}=Y_{i}-Z_{i}, i=1,2, \ldots\) If \(\alpha_{a}=\beta_{a}=0.05\), show that the test can be based upon \(\bar{X}=\bar{Y}-\bar{Z} .\) Find \(c_{0}(n)\) and \(c_{1}(n)\)
. Let \(X\) and \(Y\) have a joint bivariate normal distribution. An observation \((x, y)\) arises from the joint distribution with parameters equal to either $$\mu_{1}^{\prime}=\mu_{2}^{\prime}=0,\quad\left(\sigma_{1}^{2}\right)^{\prime}=\left(\sigma_{2}^{2}\right)^{\prime}=1, \quad \rho^{\prime}=\frac{1}{2}$$ or $$\mu_{1}^{\prime \prime}=\mu_{2}^{\prime \prime}=1,\left(\sigma_{1}^{2}\right)^{\prime \prime}=4, \quad\left(\sigma_{2}^{2}\right)^{\prime \prime}=9, \rho^{\prime \prime}=\frac{1}{2}$$ Show that the classification rule involves a second-degree polynomial in \(x\) and \(y\).
What do you think about this solution?
We value your feedback to improve our textbook solutions.