% 302f20Assignment11.tex \documentclass[12pt]{article} %\usepackage{amsbsy} % for \boldsymbol and \pmb \usepackage{graphicx} % To include pdf files! \usepackage{amsmath} \usepackage{amsbsy} \usepackage{amsfonts} \usepackage{comment} \usepackage{euscript} % for \EuScript \usepackage[colorlinks=true, pdfstartview=FitV, linkcolor=blue, citecolor=blue, urlcolor=blue]{hyperref} % For links \usepackage{fullpage} %\pagestyle{empty} % No page numbers \begin{document} %\enlargethispage*{1000 pt} \begin{center} {\Large \textbf{STA 302f20 Assignment Eleven}\footnote{This assignment was prepared by \href{http://www.utstat.toronto.edu/~brunner}{Jerry Brunner}, Department of Statistical Sciences, University of Toronto. It is licensed under a \href{http://creativecommons.org/licenses/by-sa/3.0/deed.en_US} {Creative Commons Attribution - ShareAlike 3.0 Unported License}. Use any part of it as you like and share the result freely. The \LaTeX~source code is available from the course website: \href{http://www.utstat.toronto.edu/~brunner/oldclass/302f20} {\small\texttt{http://www.utstat.toronto.edu/$^\sim$brunner/oldclass/302f20}}} } \vspace{1 mm} \end{center} \noindent The following problems are not to be handed in. They are preparation for the quiz in tutorial and the final exam. Please try them before looking at the answers. Use the formula sheet. Be ready for R questions similar to the ones asked in this assignment. You will not be asked to hand in your complete answers to the R parts of this assignment, but you might be asked to do something similar on the quiz. \begin{enumerate} %%%%%%%%%%%%%%%%%%%%%%%%% Generalized Least Squares %%%%%%%%%%%%%%%%%%%% \item \label{gls} In \emph{generalized least squares}, the regression model is $\mathbf{y} = \mathbf{X} \boldsymbol{\beta} + \boldsymbol{\epsilon}$, where $\boldsymbol{\epsilon} \sim N_n(\mathbf{0},\sigma^2\mathbf{V})$. The matrix $\mathbf{V}$ is a \emph{known} symmetric positive definite matrix. \begin{enumerate} \item Let $\widehat{\boldsymbol{\beta}}$ denote the ordinary least squares estimate. Is $\widehat{\boldsymbol{\beta}}$ still an unbiased estimator of $\boldsymbol{\beta}$ for this problem? \item What is $cov(\widehat{\boldsymbol{\beta}})$ for this problem? \item Multiply $\mathbf{y} = \mathbf{X} \boldsymbol{\beta} + \boldsymbol{\epsilon}$ on the left by $\mathbf{V}^{-1/2}$, obtaining $\mathbf{y}^* = \mathbf{X}^* \boldsymbol{\beta} + \boldsymbol{\epsilon}^*$. What is the distribution of $\boldsymbol{\epsilon}^*$? \item Substituting $\mathbf{X}^*$ and $\mathbf{y}^*$ into the usual formula for $\widehat{\boldsymbol{\beta}}$, obtain the generalized least squares estimate on page 165 of the textbook. The textbook just calls it $\widehat{\boldsymbol{\beta}}$, but we will call it $\widehat{\boldsymbol{\beta}}_{gls}$. \item What is the distribution of $\widehat{\boldsymbol{\beta}}_{gls}$? Show your work. \item Suppose you guessed wrong about $\mathbf{V}$, and the true distribution of $\boldsymbol{\epsilon}$ is $N_n(\mathbf{0},\sigma^2\boldsymbol{\Omega})$, where $\boldsymbol{\Omega}$ is symmetric and positive definite. Is $\widehat{\boldsymbol{\beta}}_{gls}$ still an unbiased estimator of $\boldsymbol{\beta}$? \item In terms of the parameters of the original model, please express the following in terms of the original model. \begin{enumerate} \item $\mathbf{H}^*$. Is $\mathbf{H}^*$ symmetric and idempotent? \item $\widehat{\mathbf{y}}^*$. \item $\widehat{\boldsymbol{\epsilon}}^*$. \item Show that \emph{SSE}$^*$ = $(\mathbf{y}-\mathbf{X}\widehat{\boldsymbol{\beta}}_{gls})^\prime \mathbf{V}^{-1} (\mathbf{y}-\mathbf{X}\widehat{\boldsymbol{\beta}}_{gls})$. % Numerator of (7.65) on p. 165 \item Write the $F$ statistic for the general linear test of $H_0: \mathbf{C}\boldsymbol{\beta} = \mathbf{t}$. Use the notation $\widehat{\boldsymbol{\beta}}_{gls}$; otherwise it's too ugly. \end{enumerate} \end{enumerate} % End of GLS \newpage %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \item \label{wlsintro} In \emph{weighted least squares}, the model is the same as in Question~\ref{gls}, except that the matrix $\mathbf{V}$ is diagonal. In scalar form, $y_i = \beta_0 + \beta_1 x_{i,1} + \cdots + \beta_k x_{i,k} + \epsilon_i$, where the $\epsilon_i$ are independent $N(0,\sigma^2v_i)$. The constants $v_1, \ldots, v_n$ are known, and strictly greater than zero. As usual, $\sigma^2$ is unknown. This means that the variances of the $y_i$ are unequal, but known up to a proportionality constant. \begin{enumerate} \item What is the distribution of $y_i$? \item Multiply both sides of the regression equation by $\frac{1}{\sqrt{v_i}}$, obtaining $y_i^* = \beta_0 x_{i,0}^* + \beta_1 x_{i,1}^* + \cdots + \beta_k x_{i,k}^* + \epsilon_i^*$. \begin{enumerate} \item What is $x_{i,0}^*$? \item What is the joint distribution of $\epsilon_1^*, \ldots, \epsilon_n^*$? \item What is the distribution of $y_i^*$? Are they independent? \end{enumerate} \item \label{help} The R help file for \texttt{lm} claims that the weighted least squares estimates of $\beta_0, \ldots, \beta_k$ are obtained by minimizing \verb!sum(w*e^2)!. In our notation, they mean minimizing $\sum_{i=1}^n w_i \widehat{\epsilon}_i^{\,2}$ over $\widehat{\beta}_0, \ldots, \widehat{\beta}_k$. Show that this yields the least squares estimates for our starred model. In our notation, what are the ``weights" $w_1, \ldots, w_n$? \end{enumerate} % End of WLSintro \item In lecture, there was an example in which $y_{ij} \stackrel{i.i.d.}{\sim} ~ ?(\mu,\sigma^2)$, and we observe $\overline{y}_1, \ldots, \overline{y}_m$ along with $n_1, \ldots, n_m$. Verify that the estimator of $\mu$ from lecture, $\frac{\sum_{j=1}^m n_j \overline{y}_j}{\sum_{j=1}^m n_j}$, is equal to $\widehat{\boldsymbol{\beta}}_{gls}$. \item \label{xsq} Independently for $i=1, \ldots, n$, let $y_i = \beta x_i + \epsilon_i$, where $\epsilon_i \sim N(0,\sigma^2x_i^2)$. \begin{enumerate} \item Based on the approach of Problem~\ref{wlsintro}, obtain the weighted least squares estimate of $\beta$. The answer is a formula for computing $\widehat{\beta}$. There's a way that uses calculus (minimizing the quantity of Question~\ref{help}), and another way that relies on known results. Either one is fine. \item Verify that the answer is a special case of $\widehat{\boldsymbol{\beta}}_{gls}$. \end{enumerate} \item Data from the example of Question~\ref{xsq} are available \href{http://www.utstat.toronto.edu/~brunner/data/legal/TinyWLS.data.txt}{here}. The URL is \\ \verb!http://www.utstat.toronto.edu/~brunner/data/legal/TinyWLS.data.txt! . \begin{enumerate} \item Estimate $\beta$ by ordinary least squares, which is the default for the \texttt{lm()} function. What is $\widehat{\beta}$? Can you reject $H_0: \beta=0$ at $\alpha = 0.05$? \item Now estimate $\beta$ by weighted least squares. What is $\widehat{\beta}_{wls}$? This time, can you reject $H_0: \beta=0$ at $\alpha = 0.05$? \item Verify that $\widehat{\beta}_{wls}$ agrees with your answer to Question \ref{xsq}. \end{enumerate} \newpage %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%%%%%%%%%%%%%% Centered model %%%%%%%%%%%%%%%%%%%% % \vspace{3mm} \hrule \vspace{3mm} %%%%%%%%%%%%%%%%%%%%%%%%%%%% \item In the \emph{centered} linear regression model, sample means are subtracted from the explanatory variables, so that values above average are positive and values below average are negative. Here is a version with one explanatory variable. The uncentered model for $y_i$ equals the centered model for $y_i$, as follows. \begin{eqnarray*} y_i & = & \beta_0 + \beta_1x_i + \epsilon_i \\ & = & \alpha_0 + \alpha_1 (x_i-\overline{x}) + \epsilon_i \end{eqnarray*} \begin{enumerate} \item Give $\alpha_0$ and $\alpha_1$ in terms of $\beta_0$ and $\beta_1$. Show your work. \item What does the $\mathbf{X}$ matrix look like for the uncentered model? What does the corresponding matrix look like for the centered model? \item Centering can be accomplished by matrix multiplication. For this simple regression example, give the matrix $\mathbf{A} = (a_{ij})$ such that $\mathbf{XA} = \mathbf{W}$, where $\mathbf{W}$ is the explanatory variable matrix for the centered model. \item Verify that \begin{displaymath} \left(\begin{array}{c} \alpha_0 \\ \alpha_1 \end{array}\right) = \left(\begin{array}{cc} a_{11} & a_{12} \\ a_{21} & a_{22} \end{array}\right)^{-1} \left(\begin{array}{c} \beta_0 \\ \beta_1 \end{array}\right). \end{displaymath} \end{enumerate} \item Consider the centered multiple regression model \begin{displaymath} y_i = \beta_0 + \beta_1 (x_{i,1}-\overline{x}_1) + \cdots + \beta_k (x_{i,k}-\overline{x}_k) + \epsilon_i \end{displaymath} with the usual details. Find the least squares estimate of $\beta_0$ by differentiating; show your work. Don't forget the second derivative test. \item For the general regression model with $k$ predictor variables, centering is also a one-to-one linear transformation. In general, \begin{eqnarray*} & & \mathbf{y} = \mathbf{X} \boldsymbol{\beta} + \boldsymbol{\epsilon} \\ & \iff & \mathbf{y} = \mathbf{XA A}^{-1} \boldsymbol{\beta} + \boldsymbol{\epsilon} \\ & \iff & \mathbf{y} = \mathbf{W} \boldsymbol{\alpha} + \boldsymbol{\epsilon}, \end{eqnarray*} where $\mathbf{A}$ is a $(k+1) \times (k+1)$ matrix, $\mathbf{W=XA}$ and $\boldsymbol{\alpha} = \mathbf{A}^{-1} \boldsymbol{\beta}$. \begin{enumerate} \item Denoting the least-squares estimate of $\boldsymbol{\alpha}$ by $\widehat{\boldsymbol{\alpha}}$, find a formula for $\widehat{\boldsymbol{\alpha}}$. Simplify. What is its connection to $\widehat{\boldsymbol{\beta}}$? \item What is the vector of predicted $y$ values for the transformed model? How does it compare to $\widehat{\mathbf{y}}$ from the original model? \item Give a null hypothesis equivalent to $H_0: \mathbf{C}\boldsymbol{\beta}=\mathbf{t}$, but in terms of the transformed model. It's $H_0: \mathbf{C}_2\boldsymbol{\alpha}=\mathbf{t}$. What is $\mathbf{C}_2$? \item Compare the $F^*$ statistics for testing $H_0: \mathbf{C}\boldsymbol{\beta}=\mathbf{t}$ and $H_0: \mathbf{C}_2\boldsymbol{\alpha}=\mathbf{t}$. One would hope they are the same. Are they? Show your work. \end{enumerate} % \vspace{30mm} \hrule \vspace{30mm} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \item This question will re-use the \texttt{sales} data from Assignment Nine. For reference, here is the description of the data again. Telephone sales representatives use computer software to help them locate potential customers, answer questions, take credit card information and place orders. Twelve sales representatives were randomly assigned to each of three new software packages the company was thinking of purchasing. The data for each sales representative include the software package (1, 2 or 3), sales last quarter with the old software, and sales this quarter with one of the new software packages. Sales are in number of units sold. The response variable is sales this quarter. The data are available \href{http://www.utstat.toronto.edu/~brunner/data/legal/sales.data.txt} {here}. The URL is \begin{center} \verb+http://www.utstat.toronto.edu/~brunner/data/legal/sales.data.txt+ \end{center} The beautiful plot on slide 13 of lecture unit 22 (The Centered Model) is actually a plot of these data. The effect of software package on sales this quarter certainly seems to depend on sales last quarter. So, all the regression models for this question should include interactions. \begin{enumerate} \item Fit an uncentered model. Test for differences in expected sales for the three software packages for sales representatives whose sales last quarter were average -- that is, exactly at the overall sample mean. Obtain the $F$ statistic, the $p$-value, and so on. What do you conclude? \item Now carry out the same test on a centered model in which Sales Last Quarter is transformed by subtracting off the sample mean. Note that the product terms in your model ar products of the dummy variable and \emph{centered} sales last quarter. Is the $F$ statistic equal to what you got for the uncentered model? \item The coloured scatterplot suggests that Software Package Two might be the best for sales reps with below average performance last quarter, and worst for sales reps with above average performance last quarter. But is it significant? Accordingly, test for differences among the three packages when Sales Last Quarter~=~85. You have to admit that nothing in the theory depends on the centering being done at exactly $\overline{x}$, so do this job by ``centering" Sales Last Quarter at 85. What do you conclude? \item Two of the three pairwise comparisons at $x=85$ are available from the output of \texttt{summary}. Carry out the third pairwise comparison. With a Bonferroni correction for all pairwise comparisons and in plain, non-statistical language, what do you conclude? \end{enumerate} % What do you conclude? \end{enumerate} % End of all the questions \end{document} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % \vspace{30mm} \hrule \vspace{30mm} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % \vspace{30mm}\hrule\vspace{30mm} \begin{comment} Dinky $Var(\epsilon_i) = \sigma^2 x_i^2$ with R Software training again with centering: R Centering with no intercept -- see book. Forget it; it's pre-multiplication. How about standardizing \end{comment} \begin{comment} # Dinky $Var(\epsilon_i) = \sigma^2 x_i^2$ with R rm(list=ls()) beta = 0.25; sigmasq = 1; n = 100 set.seed(9999) x = round(runif(n,1,5),2); epsilon = rnorm(n,0,sqrt(sigmasq*x^2)) y = beta*x + epsilon; y = round(y,2); plot(x,y) # First fit an ordinary least squares model mod1 = lm(y ~ 0 + x); summary(mod1) # Now weighted least squares mod2 = lm(y ~ 0 + x, weights = 1/x^2); summary(mod2) # Finally, verify that the estimate is just mean(y/x), as in Problem 4 mean(y/x) \end{comment}