% This version is a significant improvement on STA431s17 % \documentclass[serif]{beamer} % Serif for Computer Modern math font. \documentclass[serif, handout]{beamer} % Handout mode to ignore pause statements \hypersetup{colorlinks,linkcolor=,urlcolor=red} \usefonttheme{serif} % Looks like Computer Modern for non-math text -- nice! \setbeamertemplate{navigation symbols}{} % Suppress navigation symbols % \usetheme{Berlin} % Displays sections on top % \usetheme{Frankfurt} % Displays section titles on top: Fairly thin but still swallows some material at bottom of crowded slides % \usetheme{Berkeley} \usetheme{AnnArbor} % CambridgeUS % I'm using this one (yellow) just to be different from Dehan. \usepackage[english]{babel} \usepackage{amsmath} % for binom \usepackage{comment} % \usepackage{graphicx} % To include pdf files! % \definecolor{links}{HTML}{2A1B81} % \definecolor{links}{red} \setbeamertemplate{footline}[frame number] \mode \title{Confirmatory Factor Analysis Part Two\footnote{See last slide for copyright information.}} \subtitle{STA2053 Fall 2022} \date{} % To suppress date \begin{document} \begin{frame} \titlepage \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Original and Surrogate Models} %\framesubtitle{} \begin{itemize} \item Original model has expected values, intercepts, and slopes that need not equal one. \item Re-parameterization via a change of variables yields a surrogate model. \item Centered surrogate model has the same covariance matrix as the original. % \item If a model has any factor loadings or variances equal to one, it's a surrogate model. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Why should the variance of the factors equal one?} \pause %\framesubtitle{} \begin{itemize} \item Inherited from exploratory factor analysis, which was mostly a disaster. \item The standard answer is something like this: ``Because it’s arbitrary. The variance depends upon the scale on which the variable is measured, but we can't see it to measure it directly. So set it to one for convenience." \pause \item But saying it does not make it so. If $F$ is a random variable with an unknown variance, then \item $Var(F)=\phi$ is an unknown parameter. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Centered Model} \framesubtitle{One factor, four observed variables} \begin{columns} \column{0.5\textwidth} \begin{eqnarray*} d_1 &=& \lambda_1 F + e_1 \\ d_2 &=& \lambda_2 F + e_2 \\ d_3 &=& \lambda_3 F + e_3 \\ d_4 &=& \lambda_4 F + e_4 % 36 \end{eqnarray*} \column{0.5\textwidth} $\begin{array}{l} e_1,\ldots, e_4, F \mbox{ all independent} \\ Var(e_j) = \omega_j ~~~~ {\color{red}Var(F) = \phi} \\ \lambda_1, \lambda_2, \lambda_3 \neq 0 \end{array} $ \end{columns} \pause \vspace{10mm} % Taken from the old Sage job Fac1 on Big Iron $\boldsymbol{\Sigma} = \left(\begin{array}{rrrr} \lambda_{1}^{2} \phi + \omega_{1} & \lambda_{1} \lambda_{2} \phi & \lambda_{1} \lambda_{3} \phi & \lambda_{1} \lambda_{4} \phi \\ \lambda_{1} \lambda_{2} \phi & \lambda_{2}^{2} \phi + \omega_{2} & \lambda_{2} \lambda_{3} \phi & \lambda_{2} \lambda_{4} \phi \\ \lambda_{1} \lambda_{3} \phi & \lambda_{2} \lambda_{3} \phi & \lambda_{3}^{2} \phi + \omega_{3} & \lambda_{3} \lambda_{4} \phi \\ \lambda_{1} \lambda_{4} \phi & \lambda_{2} \lambda_{4} \phi & \lambda_{3} \lambda_{4} \phi & \lambda_{4}^{2} \phi + \omega_{4} \end{array}\right) $ \vspace{3mm} Passes the Counting Rule test with 10 equations in 9 unknowns \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{But for any $c \neq 0$} %\framesubtitle{} \begin{displaymath} \begin{array}{c|ccccccccc} \boldsymbol{\theta}_1 & \phi & \lambda_1 &\lambda_2 &\lambda_3 &\lambda_4 & \omega_1 & \omega_2 & \omega_3 & \omega_4 \\ \hline \boldsymbol{\theta}_2 & \phi/c^2 & c\lambda_1 &c\lambda_2 &c\lambda_3 & c\lambda_4 & \omega_1 & \omega_2 & \omega_3 & \omega_4 \end{array} \end{displaymath} \pause Both yield \begin{displaymath} \boldsymbol{\Sigma} = \left(\begin{array}{rrrr} \lambda_{1}^{2} \phi + \omega_{1} & \lambda_{1} \lambda_{2} \phi & \lambda_{1} \lambda_{3} \phi & \lambda_{1} \lambda_{4} \phi \\ \lambda_{1} \lambda_{2} \phi & \lambda_{2}^{2} \phi + \omega_{2} & \lambda_{2} \lambda_{3} \phi & \lambda_{2} \lambda_{4} \phi \\ \lambda_{1} \lambda_{3} \phi & \lambda_{2} \lambda_{3} \phi & \lambda_{3}^{2} \phi + \omega_{3} & \lambda_{3} \lambda_{4} \phi \\ \lambda_{1} \lambda_{4} \phi & \lambda_{2} \lambda_{4} \phi & \lambda_{3} \lambda_{4} \phi & \lambda_{4}^{2} \phi + \omega_{4} \end{array}\right) \end{displaymath} \pause The choice $\phi=1$ just sets $c=\sqrt{\phi}$: convenient but seemingly arbitrary. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Lack of identifiability} %\framesubtitle{} \begin{itemize} \item For any set of true parameter values, there are infinitely many untrue sets of parameter values that yield the same $\boldsymbol{\Sigma}$ and hence the same probability distribution of the observable data (assuming multivariate normality). \item There is no way to know the full truth based on the data, no matter how large the sample size. \item But there is a way to know the partial truth. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Certain \emph{functions} of the parameter vector are identifiable} %\framesubtitle{} At points in the parameter space where $\lambda_1, \lambda_2, \lambda_3 \neq 0$, \begin{itemize} \item $\frac{\sigma_{12}\sigma_{13}}{\sigma_{23}} = \frac{\lambda_1\lambda_2\phi\lambda_1\lambda_3\phi} {\lambda_2\lambda_3\phi} = \lambda_1^2\phi$ \item And so if $\lambda_1>0$, the function $\lambda_j\phi^{1/2}$ is identifiable \newline for $j = 1, \ldots, 4$. \pause \item $\sigma_{11} - \frac{\sigma_{12}\sigma_{13}}{\sigma_{23}} = \omega_1$, and so $\omega_j$ is identifiable for $j = 1, \ldots, 4$. \pause \item $\frac{\sigma_{13}}{\sigma_{23}} = \frac{\lambda_1\lambda_3\phi}{\lambda_2\lambda_3\phi} = \frac{\lambda_1}{\lambda_2}$, so \emph{ratios} of factor loadings are identifiable. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Reliability} %\framesubtitle{} \begin{itemize} \item Reliability is the squared correlation between the observed score and the true score. \item The proportion of variance in the observed score that is not error. \pause \item For $D_1 = \lambda_1F + e_1$ it's \end{itemize} \begin{eqnarray*} \rho^2 & = & \left( \frac{Cov(D_1,F)}{SD(D_1)SD(F)}\right)^2 \\ \pause & = & \left( \frac{\lambda_1\phi}{\sqrt{\lambda_1^2\phi+\omega_1}\sqrt{\phi}}\right)^2 \\ \pause & = & \frac{\lambda_1^2\phi}{\lambda_1^2\phi+\omega_1} \end{eqnarray*} \pause $\lambda_1^2\phi$ and $\omega_1$ are both identifiable, so we've got it. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{For completeness} %\framesubtitle{} {\small \begin{displaymath} \rho^2 = \frac{\lambda_1^2\phi}{\lambda_1^2\phi+\omega_1} \hspace{5mm} \boldsymbol{\Sigma} = \left(\begin{array}{rrrr} \lambda_{1}^{2} \phi + \omega_{1} & \lambda_{1} \lambda_{2} \phi & \lambda_{1} \lambda_{3} \phi & \lambda_{1} \lambda_{4} \phi \\ \lambda_{1} \lambda_{2} \phi & \lambda_{2}^{2} \phi + \omega_{2} & \lambda_{2} \lambda_{3} \phi & \lambda_{2} \lambda_{4} \phi \\ \lambda_{1} \lambda_{3} \phi & \lambda_{2} \lambda_{3} \phi & \lambda_{3}^{2} \phi + \omega_{3} & \lambda_{3} \lambda_{4} \phi \\ \lambda_{1} \lambda_{4} \phi & \lambda_{2} \lambda_{4} \phi & \lambda_{3} \lambda_{4} \phi & \lambda_{4}^{2} \phi + \omega_{4} \end{array}\right) \end{displaymath} \pause } % End size \begin{eqnarray*} \frac{\sigma_{12}\sigma_{13}}{\sigma_{23}\sigma_{11}} & = & \frac{\lambda_1\lambda_2\phi\lambda_1\lambda_3\phi} {\lambda_2\lambda_3\phi (\lambda_1^2\phi+\omega_1)} \\ \pause & = & \frac{\lambda_1^2\phi}{\lambda_1^2\phi+\omega_1} \\ & = & \rho^2 \end{eqnarray*} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{What can we successfully estimate?} %\framesubtitle{} \begin{itemize} \item Error variances are knowable. \item Factor loadings and variance of the factor are not knowable separately. \item But both are knowable up to multiplication by a non-zero constant, so signs of factor loadings are knowable (if one sign is known). \item Relative magnitudes (ratios) of factor loadings are knowable. \item Reliabilities are knowable. \end{itemize} \end{frame} % Had a slide here called testing model fit. It claimed that the constraints on the covariances are the same for the original model. That's true for a 1-factor model, but now I'mm not so sure about the general case. See old Powerpoint slides for this one. %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Re-parameterization} %\framesubtitle{} \begin{itemize} \item The choice $\phi=1$ is a very smart re-parameterization. \item It re-expresses the factor loadings as multiples of the square root of $\phi$. \item That is, in standard deviation units. \pause \item It preserves what information is accessible about the parameters of the original model. \item Much better than exploratory factor analysis, which lost even the signs of the factor loadings. \item This is the second major re-parameterization. The first was losing the the means and intercepts. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Re-parameterizations} %\framesubtitle{} %{\large \begin{center} Original model $\rightarrow$ Surrogate model 1 $\rightarrow$ Surrogate model 2 \ldots \end{center} %} % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Add a factor to the centered original model} %\framesubtitle{} \begin{center} \includegraphics[width=3in]{TwoFactors} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Model Equations} %\framesubtitle{} \begin{eqnarray*} d_1 &=& \lambda_1 F_1 + e_1 \\ d_2 &=& \lambda_2 F_1 + e_2 \\ d_3 &=& \lambda_3 F_1 + e_3 \\ d_4 &=& \lambda_4 F_2 + e_4 \\ d_5 &=& \lambda_5 F_2 + e_5 \\ d_6 &=& \lambda_6 F_2 + e_6 \end{eqnarray*} \vspace{5mm} {\footnotesize \begin{displaymath} cov\left( \begin{array}{c} F_1 \\ F_2 \end{array} \right) = \left( \begin{array}{c c} \phi_{11} & \phi_{12} \\ \phi_{12} & \phi_{22} \\ \end{array} \right) ~ \begin{array}{l} e_1,\ldots,e_6 \mbox{ independent of each other and of } F_1, F_2 \\ \lambda_1, \ldots \lambda_6 \neq 0 \\ Var(e_j) = \omega_j \\ \end{array} \end{displaymath} } % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Parameters are not identifiable} %\framesubtitle{} $\boldsymbol{\Sigma} = $ \\ \vspace{2mm} {\scriptsize $\left(\begin{array}{rrrrrr} \lambda_{1}^{2} \phi_{11} + \omega_{1} & \lambda_{1} \lambda_{2} \phi_{11} & \lambda_{1} \lambda_{3} \phi_{11} & \lambda_{1} \lambda_{4} \phi_{12} & \lambda_{1} \lambda_{5} \phi_{12} & \lambda_{1} \lambda_{6} \phi_{12} \\ \lambda_{1} \lambda_{2} \phi_{11} & \lambda_{2}^{2} \phi_{11} + \omega_{2} & \lambda_{2} \lambda_{3} \phi_{11} & \lambda_{2} \lambda_{4} \phi_{12} & \lambda_{2} \lambda_{5} \phi_{12} & \lambda_{2} \lambda_{6} \phi_{12} \\ \lambda_{1} \lambda_{3} \phi_{11} & \lambda_{2} \lambda_{3} \phi_{11} & \lambda_{3}^{2} \phi_{11} + \omega_{3} & \lambda_{3} \lambda_{4} \phi_{12} & \lambda_{3} \lambda_{5} \phi_{12} & \lambda_{3} \lambda_{6} \phi_{12} \\ \lambda_{1} \lambda_{4} \phi_{12} & \lambda_{2} \lambda_{4} \phi_{12} & \lambda_{3} \lambda_{4} \phi_{12} & \lambda_{4}^{2} \phi_{22} + \omega_{4} & \lambda_{4} \lambda_{5} \phi_{22} & \lambda_{4} \lambda_{6} \phi_{22} \\ \lambda_{1} \lambda_{5} \phi_{12} & \lambda_{2} \lambda_{5} \phi_{12} & \lambda_{3} \lambda_{5} \phi_{12} & \lambda_{4} \lambda_{5} \phi_{22} & \lambda_{5}^{2} \phi_{22} + \omega_{5} & \lambda_{5} \lambda_{6} \phi_{22} \\ \lambda_{1} \lambda_{6} \phi_{12} & \lambda_{2} \lambda_{6} \phi_{12} & \lambda_{3} \lambda_{6} \phi_{12} & \lambda_{4} \lambda_{6} \phi_{22} & \lambda_{5} \lambda_{6} \phi_{22} & \lambda_{6}^{2} \phi_{22} + \omega_{6} \end{array}\right)$ } % End size \begin{eqnarray*} \boldsymbol{\theta}_1 &=& (\lambda_1, \ldots, \lambda_6, \phi_{11}, \phi_{12}, \phi_{22}, \omega_1, \ldots, \omega_6) \\ \boldsymbol{\theta}_2 &=& (\lambda_1^\prime, \ldots, \lambda_6^\prime, \phi_{11}^\prime, \phi_{12}^\prime, \phi_{22}^\prime, \omega_1^\prime, \ldots, \omega_6^\prime) \end{eqnarray*} $\begin{array}{rrrr} \lambda_1^\prime = c_1\lambda_1 & \lambda_2^\prime = c_1\lambda_2 & \lambda_3^\prime = c_1\lambda_3 & \phi_{11}^\prime = \phi_{11}/c_1^2 \\ \lambda_4^\prime = c_2\lambda_4 & \lambda_5^\prime = c_2\lambda_5 & \lambda_6^\prime = c_2\lambda_6 & \phi_{22}^\prime = \phi_{22}/c_2^2 \\ \phi_{12}^\prime = \frac{\phi_{12}}{c_1c_2} \end{array} $ \\ ~$\omega_j^\prime = \omega_j \mbox{ for } j = 1, \ldots, 6$ \end{frame} \begin{comment} --------------------------------------------------------------------- SageMath code load sem L = ZeroMatrix(6,2) L[0,0]= var('lambda1'); L[1,0]= var('lambda2'); L[2,0]= var('lambda3') L[3,1]= var('lambda4'); L[4,1]= var('lambda5'); L[5,1]= var('lambda6') P = SymmetricMatrix(2,'phi') O = DiagonalMatrix(6,symbol='omega') Sig = FactorAnalysisVar(L,P,O); Sig Sig=expand(Sig); Sig print(latex(Sig)) \end{comment} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Variances and covariances of factors} %\framesubtitle{} \begin{itemize} \item Are knowable only up to multiplication by positive constants. \pause \item Since the parameters of the latent variable model will be recovered from $\boldsymbol{\Phi} = cov(\mathbf{F})$, they also will be knowable only up to multiplication by positive constants -- at best. \pause \item Luckily, in most applications the interest is in testing (pos-neg-zero) more than estimation. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{$cov(F_1,F_2)$ is un-knowable, but} %\framesubtitle{} \begin{itemize} \item Easy to tell if it’s zero. \item Sign is known if one factor loading from each set is known -- say $\lambda_1>0$, $\lambda_4>0$. \pause \item And, \begin{eqnarray*} \frac{\sigma_{14}} {\sqrt{\frac{\sigma_{12}\sigma_{13}}{\sigma_{23}}} \sqrt{\frac{\sigma_{45}\sigma_{46}}{\sigma_{56}}}} &=& \frac{\lambda_1\lambda_4\phi_{12}} {\lambda_1\sqrt{\phi_{11}}\lambda_4\sqrt{\phi_{22}}} \\ \pause &=& \frac{\phi_{12}}{\sqrt{\phi_{11}}\sqrt{\phi_{22}}} \\ \pause &=& Corr(F_1,F_2) \end{eqnarray*} \item The \emph{correlation} between factors is identifiable! \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{The correlation between factors is identifiable} %\framesubtitle{} \begin{itemize} \item Furthermore, it is the same function of $\boldsymbol{\Sigma}$ that yields $\phi_{12}$ under the surrogate model with $Var(F_1) = Var(F_2) = 1$. \pause \item Therefore, $Corr(F_1,F_2) = \phi_{12}$ under the surrogate model is actually $Corr(F_1,F_2)$ under the original model. \pause \item Estimation is very meaningful. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Setting variances of factors to one} %\framesubtitle{} \begin{itemize} \item Is a \emph{very} smart re-parameterization. \item Is excellent when the interest is in correlations between factors. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Re-parameterization as a change of variables} %\framesubtitle{} {\LARGE \begin{eqnarray*} d_j &=& \lambda_j F_j + e_j \\ \pause &=& (\lambda_j \sqrt{\phi_{jj}}) \left(\frac{1}{\sqrt{\phi_{jj}}}F_j\right) + e_j \\ \pause &=& \lambda_j^\prime F_j^\prime + e_j \end{eqnarray*} } % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Covariances} %\framesubtitle{} {\large \begin{eqnarray*} Cov(F^\prime_j, F^\prime_k) &=& Cov\left(\frac{1}{\sqrt{\phi_{jj}}}F_j, \frac{1}{\sqrt{\phi_{kk}}}F_k\right) \\ \pause &=& \frac{Cov(F_j,F_k)}{\sqrt{\phi_{jj}}\sqrt{\phi_{kk}}} \\ \pause &=& \frac{\phi_{jk}}{\sqrt{\phi_{jj}}\sqrt{\phi_{kk}}} \\ &&\\ &=& Corr(F_j,F_k) \end{eqnarray*} % 24\end{eqnarray* } % End size \end{frame} % Omitted slides: What happens if there is a latent variable model? %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Cascading effects} %\framesubtitle{} \begin{itemize} \item Understand the re-parameterization as a change of variables. \item Not just an arbitrary restriction of the parameter space. \item It shows there are widespread effects throughout the model. \item Especially if there is a detailed latent variable model. \item Also shows \emph{how} the meanings of other model parameters are affected. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{The other standard trick} %\framesubtitle{} \begin{itemize} \item Setting variances of all the factors to one is an excellent re-parameterization in disguise. \item The other standard trick is to set a factor loading equal to one for each factor. \pause \item $d = F + e$ is hard to believe if you take it literally. \item It's actually a re-parameterization. \item Every model you've seen with a factor loading of one is a surrogate model. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Back to a single-factor model with $\lambda_1>0$} %\framesubtitle{} \begin{columns} \column{0.5\textwidth} \begin{eqnarray*} d_1 &=& \lambda_1 F + e_1 \\ d_2 &=& \lambda_2 F + e_2 \\ d_3 &=& \lambda_3 F + e_3 \\ & \vdots & \end{eqnarray*} \pause \column{0.5\textwidth} \begin{eqnarray*} d_j &=& \left( \frac{\lambda_j}{\lambda_1} \right) (\lambda_1 F) + e_j \\ &=& \lambda_j^\prime F^\prime + e_j % 32 \end{eqnarray*} \pause \end{columns} \vspace{5mm} \begin{eqnarray*} d_1 &=& F^\prime + e_1 \\ d_2 &=& \lambda_2^\prime F^\prime + e_2 \\ d_3 &=& \lambda_3^\prime F^\prime + e_3 \\ & \vdots & \end{eqnarray*} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{$\boldsymbol{\Sigma}$ under the surrogate model} \framesubtitle{Covariance structure equations} \begin{displaymath} \boldsymbol{\Sigma} = % Edited \left(\begin{array}{rrr} \phi + \omega_{1} & \lambda_{2} \phi & \lambda_{3} \phi \\ \lambda_{2} \phi & \lambda_{2}^{2} \phi + \omega_{2} & \lambda_{2} \lambda_{3} \phi \\ \lambda_{3} \phi & \lambda_{2} \lambda_{3} \phi & \lambda_{3}^{2} \phi + \omega_{3} \end{array}\right) \end{displaymath} \pause % \vspace{5mm} Solutions assuming $\lambda_2$ and $\lambda_3$ non-zero: \begin{eqnarray*} \lambda_2 &=& \frac{\sigma_{23}}{\sigma_{13}} \\ \lambda_3 &=& \frac{\sigma_{23}}{\sigma_{12}} \\ \phi &=& \frac{\sigma_{12}\sigma_{13}}{\sigma_{23}} \end{eqnarray*} We have another three-variable rule. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Combined three-variable identification rule} \framesubtitle{For the record} For a centered factor analysis model with a single factor, the parameters will be identifiable provided that \begin{itemize} \item There are at least three reference variables. \item Either the factor is standardized and the sign one factor loading is known, or else at least one factor loading equals one. \item Errors are independent of one another and of the factors. \end{itemize} \pause \vspace{3mm} Add more variables and more factors using other rules. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Two Versions of $\boldsymbol{\Sigma}$} \framesubtitle{For the surrogate model with a factor loading set to one} \begin{columns} \column{0.5\textwidth} \begin{center} Under the Surrogate Model \end{center} $\left(\begin{array}{rrr} \phi + \omega_{1} & \lambda_{2} \phi & \lambda_{3} \phi \\ & \lambda_{2}^{2} \phi + \omega_{2} & \lambda_{2} \lambda_{3} \phi \\ & & \lambda_{3}^{2} \phi + \omega_{3} \end{array}\right)$ \column{0.5\textwidth} \begin{center} Under the Original Model \end{center} $\left(\begin{array}{rrr} \lambda_{1}^{2} \phi + \omega_{1} & \lambda_{1} \lambda_{2} \phi & \lambda_{1} \lambda_{3} \phi \\ & \lambda_{2}^{2} \phi + \omega_{2} & \lambda_{2} \lambda_{3} \phi \\ & & \lambda_{3}^{2} \phi + \omega_{3} \end{array}\right)$ \end{columns} \vspace{5mm} \begin{center} \begin{tabular}{|c|cc|} \hline & \multicolumn{2}{c|}{Value under model} \\ Function of $\boldsymbol{\Sigma}$ & Surrogate& Original\\ \hline $\frac{\sigma_{23}}{\sigma_{13}}$ & $\lambda_2$ & $\frac{\lambda_2}{\lambda_1}$ \\ \hline $\frac{\sigma_{23}}{\sigma_{12}}$ & $\lambda_3$ & $\frac{\lambda_3}{\lambda_1}$ \\ \hline $\frac{\sigma_{12}\sigma_{13}}{\sigma_{23}}$ & $\phi$ & $\lambda_1^2\phi$ \\ \hline \end{tabular} \end{center} \end{frame} \begin{comment} % Sage load sem L = ZeroMatrix(3,1) L[0,0]= 1; L[1,0]= var('lambda2'); L[2,0]= var('lambda3') P = ZeroMatrix(1,1); P[0,0] = var('phi'); P O = DiagonalMatrix(3,symbol='omega') Sig = FactorAnalysisVar(L,P,O); Sig print(latex(Sig)) \end{comment} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Under the surrogate model} %\framesubtitle{} \begin{itemize} \item It looks like $\lambda_j$ is identifiable, but actually it's $\lambda_j/\lambda_1$. \item Estimates of $\lambda_j$ for $j\neq 1$ are actually estimates of $\lambda_j/\lambda_1$. \item It looks like $\phi$ is identifiable, but actually it's $\lambda_1^2\phi$. \item $\phi$ is being expressed as a multiple of $\lambda_1^2$. \item Estimates of $\phi$ are actually estimates of $\lambda_1^2\phi$. \pause \item[] \item Make $d_1$ the clearest representative of the factor. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Add an observable variable to the surrogate model} %\framesubtitle{} \begin{itemize} \item Parameters are all identifiable, even if the factor loading of the new variable equals zero. \item Equality restrictions on $\boldsymbol{\Sigma}$ are created, because we are adding more equations than unknowns. \item These equality restrictions apply to the original model. \item It is straightforward to see what the restrictions are, though the calculations can be time consuming. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Finding the equality restrictions} %\framesubtitle{} \begin{itemize} \item Calculate $\boldsymbol{\Sigma}(\boldsymbol{\theta})$. \item Solve the covariance structure equations explicitly, obtaining $\boldsymbol{\theta}$ as a function of $\boldsymbol{\Sigma}$. \item Substitute the solutions back into $\boldsymbol{\Sigma}(\boldsymbol{\theta})$. \item Simplify. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Example: Add a 4th variable} %\framesubtitle{} %{\LARGE \begin{eqnarray*} d_1 &=& ~~~ F + e_1 \\ d_2 &=& \lambda_2 F + e_2 \\ d_3 &=& \lambda_3 F + e_3 \\ d_4 &=& \lambda_4 F + e_4 \end{eqnarray*} \begin{displaymath} \begin{array}{l} e_1,\ldots, e_4, F \mbox{ all independent} \\ Var(e_j) = \omega_j ~~~~ Var(F) = \phi \\ \lambda_1, \lambda_2, \lambda_3 \neq 0 \end{array} \end{displaymath} %} % End size \vspace{3mm} There are 8 parameters. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Covariance Matrix} %\framesubtitle{} \begin{displaymath} \boldsymbol{\Sigma}(\boldsymbol{\theta}) = % Edit upper left cell \left(\begin{array}{rrrr} \phi+\omega_{1} & \lambda_{2} \phi & \lambda_{3} \phi & \lambda_{4} \phi \\ \lambda_{2} \phi & \lambda_{2}^{2} \phi + \omega_{2} & \lambda_{2} \lambda_{3} \phi & \lambda_{2} \lambda_{4} \phi \\ \lambda_{3} \phi & \lambda_{2} \lambda_{3} \phi & \lambda_{3}^{2} \phi + \omega_{3} & \lambda_{3} \lambda_{4} \phi \\ \lambda_{4} \phi & \lambda_{2} \lambda_{4} \phi & \lambda_{3} \lambda_{4} \phi & \lambda_{4}^{2} \phi + \omega_{4} \end{array}\right) \end{displaymath} \pause \begin{columns} \column{0.5\textwidth} Solutions \begin{itemize} \item[]$\lambda_2 = \frac{\sigma_{23}}{\sigma_{13}}$ \item[]$\lambda_3 = \frac{\sigma_{23}}{\sigma_{12}}$ \item[]$\lambda_4 = \frac{\sigma_{24}}{\sigma_{12}}$ \item[]$\phi = \frac{\sigma_{12}\sigma_{13}}{\sigma_{23}}$ \end{itemize} \pause \column{0.5\textwidth} Substitute \begin{eqnarray*} \sigma_{12} &=& \lambda_2 \phi \\ \pause &=& \frac{\sigma_{23}}{\sigma_{13}} \frac{\sigma_{12}\sigma_{13}}{\sigma_{23}} \\ \pause &=& \sigma_{12} \end{eqnarray*} \end{columns} \end{frame} % Sagemath \begin{comment} load sem L = ZeroMatrix(4,1) L[0,0]= 1; L[1,0]= var('lambda2'); L[2,0]= var('lambda3') L[3,0]= var('lambda4') P = ZeroMatrix(1,1); P[0,0] = var('phi'); P O = DiagonalMatrix(4,symbol='omega') Sig = FactorAnalysisVar(L,P,O); Sig print(latex(Sig)) \end{comment} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Substitute solutions into expressions for the covariances} %\framesubtitle{} {\small \begin{displaymath} \boldsymbol{\Sigma}(\boldsymbol{\theta}) = \left(\begin{array}{rrrr} \phi+\omega_{1} & \lambda_{2} \phi & \lambda_{3} \phi & \lambda_{4} \phi \\ \lambda_{2} \phi & \lambda_{2}^{2} \phi + \omega_{2} & \lambda_{2} \lambda_{3} \phi & \lambda_{2} \lambda_{4} \phi \\ \lambda_{3} \phi & \lambda_{2} \lambda_{3} \phi & \lambda_{3}^{2} \phi + \omega_{3} & \lambda_{3} \lambda_{4} \phi \\ \lambda_{4} \phi & \lambda_{2} \lambda_{4} \phi & \lambda_{3} \lambda_{4} \phi & \lambda_{4}^{2} \phi + \omega_{4} \end{array}\right) \end{displaymath} } % End size \begin{columns} \column{0.5\textwidth} Solutions \begin{itemize} \item[]$\lambda_2 = \frac{\sigma_{23}}{\sigma_{13}}$ \item[]$\lambda_3 = \frac{\sigma_{23}}{\sigma_{12}}$ \item[]$\lambda_4 = \frac{\sigma_{24}}{\sigma_{12}}$ \item[]$\phi = \frac{\sigma_{12}\sigma_{13}}{\sigma_{23}}$ \end{itemize} \pause \column{0.5\textwidth} % Substitute \begin{eqnarray*} \sigma_{12} &=& \sigma_{12} \\ \sigma_{13} &=& \sigma_{13} \\ \sigma_{14} &=& \frac{\sigma_{24}\sigma_{13}}{\sigma_{23}} \\ \sigma_{23} &=& \sigma_{23} \\ \sigma_{24} &=& \sigma_{24} \\ \sigma_{34} &=& \frac{\sigma_{24}\sigma_{13}}{\sigma_{12}} \end{eqnarray*} \end{columns} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Equality Constraints} \framesubtitle{There should be 4(4+1)/2 - 8 = 2} \begin{eqnarray*} \sigma_{14}\sigma_{23} &=& \sigma_{24}\sigma_{13} \\ \sigma_{12}\sigma_{34} &=& \sigma_{24}\sigma_{13} \end{eqnarray*} \pause These hold regardless of whether factor loadings are zero (1234). \begin{displaymath} \sigma_{12}\sigma_{34} = \sigma_{13}\sigma_{24} = \sigma_{14}\sigma_{23} \end{displaymath} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Add another 3-variable factor} %\framesubtitle{} \begin{itemize} \item Identifiability is maintained. \pause \item $\sigma_{14} = Cov(d_1,d_4) \pause = Cov(F_1+e_1,F_2+e_4) \pause = Cov(F_1,F_2) = \phi_{12}$ \pause \item Actually, $\sigma_{14} = \lambda_1 \lambda_4 \phi_{12}$ under the original model. \item The covariances of the surrogate model are just those of the surrogate model, multiplied by un-knowable positive constants. \item As more variables and more factors are added, all this remains true. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Comparing the surrogate models} %\framesubtitle{} \begin{itemize} \item Either set variances of factors to one, or set one loading per factor to one. \pause \item Both arise from a similar change of variables. \item $F^\prime_j = \lambda_jF_j$ or $F^\prime_j = \frac{1}{\sqrt{\phi_{jj}}} F_j$. \pause \item \emph{Meaning} of surrogate model parameters in terms of the original model is different except for the signs. \item Both surrogate models share the same equality constraints, and hence the same goodness of fit results for any given data set. \item These constraints are also true of the original model. % I don't quite have a proof, but consider the process of solving for identifiable functions of the parameters, and then substituting back. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{The Equivalence Rule} \framesubtitle{See text for proof} For a centered factor analysis model with at least one reference variable for each factor, suppose that surrogate models are obtained by either standardizing the factors, or by setting the factor loading of a reference variable equal to one for each factor. Then the parameters of one surrogate model are identifiable if and only if the parameters of the other surrogate model are identifiable. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Which re-parameterization is better?} %\framesubtitle{} {\small \begin{itemize} \item Technically, they are equivalent. \item Interpretation of the surrogate model parameters is different. \pause % \item They both involve setting a single un-knowable parameter to one, for each factor. % \item This seems arbitrary, but actually it results in a very good re-parameterization that preserves what is knowable about the true model. \item Standardizing the factors (Surrogate model 2A) is more convenient for estimating correlations between factors. \item But it's not robust to normality, so bootstrap it. \pause \item Setting one loading per factor equal to one (Surrogate model 2B) is more convenient for estimating the relative sizes of factor loadings. \item Hand calculations and identifiability proofs with Surrogate model 2B can be easier. \item If there is a serious latent variable model, Surrogate model 2B is much easier to specify with lavaan. \item Mixing Surrogate model 2B with double measurement is natural. \pause \item Don't do both restrictions for the same factor! \end{itemize} } % End size \end{frame} % herehere page 35 %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Why are we doing this? To buy identifiability.} %\framesubtitle{} \begin{itemize} \item The parameters of the original model cannot be estimated directly. For example, maximum likelihood will fail because the maximum is not unique. \item The parameters of the surrogate models are identifiable (estimable) functions of the parameters of the true model. \pause \item They have the same signs (positive, negative or zero) as the corresponding parameters of the true model. \item Hypothesis tests mean what you think they do. \item Parameter estimates can be useful if you know what the new parameters mean. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Crossover Patterns} %\framesubtitle{} \begin{itemize} \item It is unfortunate when variables can only be infuenced by one factor. In fact, it’s unbelievable most of the time. \item A pattern like this would be nicer. \end{itemize} \begin{center} \includegraphics[width=2.9in]{RealEstate} \end{center} % This does not fit any rules but I think I can get it by setting the first blue loaing to one for each factor, and then, for example, calculating Cov(Sat1,Q1) and Cov(Sat2,Q1). \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{The Extra Variables Rule} % \framesubtitle{} A set of observable variables may be added to a measurement model whose parameters are \textbf{already identifiable}, provided \begin{itemize} \item There is a reference variable for each factor in the existing model. \item Error terms of the additional variables have zero covariance with the error terms of the reference variables in the existing model. \item Error terms of the additional variables have zero covariance with the factors. \end{itemize} \pause Under these conditions, \begin{itemize} \item Straight arrows with factor loadings on them may point from each existing factor to each new variable. \item Error terms for the new set of variables may have non-zero covariances with each other, {\color{blue}and with the error terms in the original model that do not belong to the reference variables}. \item You don't need to include all such links. \end{itemize} All parameters of the new model are identifiable. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Adding Variables} %\framesubtitle{} \begin{center} \includegraphics[width=4in]{Crossover} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Proof of The Extra Variables Rule} \framesubtitle{Extra variables are in $\mathbf{d}_3$} {\Large \begin{eqnarray*} \mathbf{d}_1 &=& \mathbf{F} + \mathbf{e}_1 \\ \mathbf{d}_2 &=& \boldsymbol{\Lambda}_2\mathbf{F} + \mathbf{e}_2 \\ \mathbf{d}_3 &=& {\color{red} \boldsymbol{\Lambda}_3} \mathbf{F} + \mathbf{e}_3 \end{eqnarray*} \vspace{4mm} \begin{displaymath} cov\left( \begin{array}{c} \mathbf{e}_1 \\ \hline \mathbf{e}_3 \\ \hline \mathbf{e}_3 \end{array} \right) = \left( \begin{array}{r|r|r} \boldsymbol{\Omega}_{11} & \boldsymbol{\Omega}_{12} & \mathbf{0}\\ \hline & \boldsymbol{\Omega}_{22} & {\color{blue}\boldsymbol{\Omega}_{23}}\\ \hline & & {\color{red}\boldsymbol{\Omega}_{33}} \end{array} \right) \end{displaymath} } % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Solving for ${\color{red}\boldsymbol{\Lambda}_3}$, ${\color{blue}\boldsymbol{\Omega}_{23}}$ and ${\color{red}\boldsymbol{\Omega}_{33}}$} \pause %\framesubtitle{} \begin{eqnarray*} \boldsymbol{\Sigma} = cov\left( \begin{array}{c} \mathbf{d}_1 \\ \hline \mathbf{d}_2 \\ \hline \mathbf{d}_3 \end{array} \right) &=& \left( \begin{array}{r|r|r} \boldsymbol{\Sigma}_{11} & \boldsymbol{\Sigma}_{12} & \boldsymbol{\Sigma}_{13} \\ \hline & \boldsymbol{\Sigma}_{22} & \boldsymbol{\Sigma}_{23} \\ \hline & & \boldsymbol{\Sigma}_{33} \end{array} \right) \\ \\ &=& \left( \begin{array}{r|r|r} \boldsymbol{\boldsymbol{\Phi}+\Omega}_{11} & \boldsymbol{\Phi}\boldsymbol{\Lambda}_2^\top & \boldsymbol{\Phi} {\color{red}\boldsymbol{\Lambda}_3^\top} \\ \hline & \boldsymbol{\Lambda}_2\boldsymbol{\Phi}\boldsymbol{\Lambda}_2^\top + \boldsymbol{\Omega}_{22} & \boldsymbol{\Lambda}_2\boldsymbol{\Phi}{\color{red}\boldsymbol{\Lambda}_3^\top} + {\color{blue}\boldsymbol{\Omega}_{23}} \\ \hline & & {\color{red}\boldsymbol{\Lambda}_3}\boldsymbol{\Phi} {\color{red}\boldsymbol{\Lambda}_3^\top} + {\color{red}\boldsymbol{\Omega}_{33}} \end{array} \right) \end{eqnarray*} \pause Solve for a parameter and it becomes black. \pause \begin{eqnarray*} {\color{red}\boldsymbol{\Lambda}_3} &=& \boldsymbol{\Sigma}_{13}^\top\boldsymbol{\Phi}^{-1} \\ \pause {\color{red}\boldsymbol{\Omega}_{33}} &=& \boldsymbol{\Sigma}_{33} - \boldsymbol{\Lambda}_3\boldsymbol{\Phi}\boldsymbol{\Lambda}_3^\top \\ \pause {\color{blue}\boldsymbol{\Omega}_{23}} &=& \boldsymbol{\Sigma}_{23} - \boldsymbol{\Lambda}_2\boldsymbol{\Phi}\boldsymbol{\Lambda}_3^\top ~~ \blacksquare \end{eqnarray*} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{It's not so easy to put in all possible blue links} %\framesubtitle{} \begin{center} \includegraphics[width=3in]{Crossover} \end{center} \pause {\footnotesize \begin{itemize} \item Error terms of $e_2$, $e_3$, $e_4$ and $e_5$, can have covariances with $e_7 \ldots e_{14}$. \pause \item Also, the Factor Model Combination Rule says $e_2$ may have covariances with $e_4$ and $e_5$, and $e_3$ may have covariances with $e_4$ and $e_5$. \pause \item But there still cannot be covariances between $e_2$ and $e_3$, or between $e_4$ and $e_5$. \pause \item It's hard to draw. \end{itemize} } % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{We can do a bit better} \framesubtitle{Adding $\mathbf{d}_3$ with the Extra Variables Rule} Instead of \begin{eqnarray*} \mathbf{d}_1 &=& \mathbf{F} + \mathbf{e}_1 \\ \mathbf{d}_2 &=& \boldsymbol{\Lambda}_2\mathbf{F} + \mathbf{e}_2 \\ \mathbf{d}_3 &=& {\color{red} \boldsymbol{\Lambda}_3} \mathbf{F} + \mathbf{e}_3 \end{eqnarray*} Could have \begin{eqnarray*} \mathbf{d}_1 &=& \boldsymbol{\Lambda}_1 \mathbf{F} + \mathbf{e}_1 \\ \mathbf{d}_2 &=& \boldsymbol{\Lambda}_2\mathbf{F} + \mathbf{e}_2 \\ \mathbf{d}_3 &=& {\color{red} \boldsymbol{\Lambda}_3} \mathbf{F} + \mathbf{e}_3, \end{eqnarray*} with the $\boldsymbol{\Lambda}_1$ matrix $p \times p$ and invertable. \end{frame} % Had some "Add an observed variable to the factors" slides, but I did that back in the general model. \begin{comment} \end{comment} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{We have some identifiability rules for the measurement model} %\framesubtitle{} \begin{itemize} \item Double Measurement rule. \item Scalar three-variable rule(s). \item The equivalence rule. \item Factor combination rule. % This version is better than the one in the text. \item Extra variable rule. \item Error-free rule. \pause \item Need the 2-variable rules. \item Need the vector 3-variable rule. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Two-variable Rule} %\framesubtitle{} The parameters of a factor analysis model are identifiable provided \begin{itemize} \item There are two factors. \item There are two reference variables for each factor. \item For each factor, either the variance equals one and the sign of one factor loading is known, or the factor loading of at least one reference variable is equal to one. \item The two factors have non-zero covariance. \item Errors are independent of one another and of the factors. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Two-variable Rule} \framesubtitle{Two factors and two reference variables per factor} \vspace{-5mm} {\LARGE \begin{eqnarray*} d_1 &=& F_1 + e_1 \\ d_2 &=& \lambda_2 F_1 + e_2 \\ d_3 &=& F_2 + e_3 \\ d_4 &=& \lambda_4 F_2 + e_4, \end{eqnarray*} } % End size \noindent with all expected values zero, $cov\left(\begin{array}{c} F_1 \\ F_2 \end{array}\right) = \boldsymbol{\Phi} = [\phi_{i j}]$, $Var(e_j)=\omega_j$, and the error terms independent of the factors and each other. An additional critical stipulation is that $Cov(F_1,F_2) = \phi_{12} \neq 0$. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Covariance matrix} %\framesubtitle{} %{\Large \begin{eqnarray*} \boldsymbol{\Sigma} & = & cov\left(\begin{array}{c} d_1 \\ d_2 \\ d_3 \\ d_4 \end{array}\right) = \left(\begin{array}{cccc} \sigma_{11} & \sigma_{12} & \sigma_{13} & \sigma_{14} \\ & \sigma_{22} & \sigma_{23} & \sigma_{24} \\ & & \sigma_{33} & \sigma_{34} \\ & & & \sigma_{44} \end{array}\right) \\ \\ & = & \left(\begin{array}{rrrr} \phi_{11} + \omega_{1} & \lambda_{2} \phi_{11} & \phi_{12} & \lambda_{4} \phi_{12} \\ & \lambda_{2}^{2} \phi_{11} + \omega_{2} & \lambda_{2} \phi_{12} & \lambda_{2} \lambda_{4} \phi_{12} \\ & & \phi_{22} + \omega_{3} & \lambda_{4} \phi_{22} \\ & & & \lambda_{4}^{2} \phi_{22} + \omega_{4} \end{array}\right) \end{eqnarray*} %} % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Two-variable Addition Rule} %\framesubtitle{} A factor with just two reference variables may be added to a measurement model whose parameters are identifiable, and the parameters of the combined model will be identifiable provided \begin{itemize} \item The errors for the two additional reference variables are independent of one another and of the error terms already in the model. \item Either the variance of the additional factor equals one and the sign of one factor loading is known, or the factor loading of at least one reference variable is equal to one. \item In the existing model with identifiable parameters, \begin{itemize} \item There is at least one reference variable for each factor, and \item At least one factor has a non-zero covariance with the new factor. \end{itemize} \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Brand Awareness Model 1} %\framesubtitle{} \begin{center} \includegraphics[width=3.5in]{Doughnut1} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Brand Awareness Model 2} %\framesubtitle{} \begin{center} \includegraphics[width=4.5in]{Doughnut2} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Brand Awareness Model 3} %\framesubtitle{} \begin{center} \includegraphics[width=3in]{Doughnut3} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Brand Awareness Model 4} %\framesubtitle{} \begin{center} \includegraphics[width=4.5in]{Doughnut4} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Brand Awareness Model 5} %\framesubtitle{} \begin{center} \includegraphics[width=3.5in]{Doughnut5} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{A big complicated measurement model} % From the text %\framesubtitle{} \begin{center} \vspace{-5mm} \includegraphics[width=3.5in]{bigCFA} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Vector 3-variable Rule} % Versions of this were called Joreskog's rule and then the reference variables rule. %\framesubtitle{} Let \begin{eqnarray*} \mathbf{d}_1 & = & \mathbf{F} + \mathbf{e}_1 \\ \mathbf{d}_2 & = & \boldsymbol{\Lambda}_2\mathbf{F} + \mathbf{e}_2 \\ \mathbf{d}_3 & = & \boldsymbol{\Lambda}_3\mathbf{F} + \mathbf{e}_3 \\ \end{eqnarray*} where \begin{itemize} \item $\mathbf{F}$, $\mathbf{d}_1$ and $\mathbf{d}_2$ and $\mathbf{d}_3$ are all $p \times 1$. \item $\boldsymbol{\Lambda}_2$ and $\boldsymbol{\Lambda}_3$ have inverses. \item $cov(\mathbf{F},\mathbf{e}_j) = cov(\mathbf{e}_i,\mathbf{e}_j) = \mathbf{O}$ \end{itemize} \pause \vspace{4mm} Then all the parameters are identifiable. \end{frame} % herehere slide 58 %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Proof of Vector 3-variable Rule} \framesubtitle{Need to identify $\boldsymbol{\Phi}$, $\boldsymbol{\Lambda}_2$, $\boldsymbol{\Lambda}_3$, $\boldsymbol{\Omega}_{11}$, $\boldsymbol{\Omega}_{22}$, $\boldsymbol{\Omega}_{33}$} \begin{columns} \column{0.3\textwidth} \begin{eqnarray*} \mathbf{d}_1 &=& \mathbf{F} + \mathbf{e}_1 \\ \mathbf{d}_2 &=& \boldsymbol{\Lambda}_2\mathbf{F} + \mathbf{e}_2 \\ \mathbf{d}_3 &=& \boldsymbol{\Lambda}_3 \mathbf{F} + \mathbf{e}_3 \end{eqnarray*} \column{0.7\textwidth} ~~~~$\boldsymbol{\Sigma} = $ \begin{displaymath} \renewcommand{\arraystretch}{1.2} \left( \begin{array}{r|r|r} \boldsymbol{\boldsymbol{\Phi}+\Omega}_{11} & \boldsymbol{\Phi}\boldsymbol{\Lambda}_2^\top & \boldsymbol{\Phi} \boldsymbol{\Lambda}_3^\top \\ \hline & \boldsymbol{\Lambda}_2\boldsymbol{\Phi}\boldsymbol{\Lambda}_2^\top + \boldsymbol{\Omega}_{22} & \boldsymbol{\Lambda}_2\boldsymbol{\Phi}\boldsymbol{\Lambda}_3^\top \\ \hline & & \boldsymbol{\Lambda}_3\boldsymbol{\Phi} \boldsymbol{\Lambda}_3^\top + \boldsymbol{\Omega}_{33} \end{array} \right) \renewcommand{\arraystretch}{1.0} \end{displaymath} \end{columns} \pause \begin{eqnarray*} \boldsymbol{\Lambda}_2 & = & \boldsymbol{\Sigma}_{23}\boldsymbol{\Sigma}_{13}^{-1} \\ \pause \boldsymbol{\Phi} & = & \boldsymbol{\Sigma}_{12}\left( \boldsymbol{\Lambda}_2^\top\right)^{-1} \\ \pause \boldsymbol{\Lambda}_3 & = & \left(\boldsymbol{\Phi}^{-1} \boldsymbol{\Sigma}_{13} \right)^\top \\ \pause \boldsymbol{\Omega}_{11},\boldsymbol{\Omega}_{22},\boldsymbol{\Omega}_{33} && \mbox{ by subtraction.} \end{eqnarray*} $\blacksquare$ \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Don't need to include all possible links between error terms} %\framesubtitle{} \noindent By \emph{omitting} certain permissible links between error terms, you can apply the Vector 3-variable Rule and the Additional Variables Rule to get some useful models that can be identified visually. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Vector 3-variable Rule} % and Additional Variables Rule %\framesubtitle{} \begin{center} \includegraphics[width=4in]{Crossover2a} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Vector 3-variable Rule and Additional Variables Rule} % %\framesubtitle{} \begin{center} \includegraphics[width=4in]{Crossover2b} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Drop some covariances between errors} % %\framesubtitle{} \begin{center} \includegraphics[width=4in]{Crossover2c} \end{center} % Exploratory factor analysis with reference variables. Note factors can be correlated. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Moral of the Story} %\framesubtitle{} With a reference variable for each factor, you can have a classical unrestricted exploratory factor analysis model for a set of additional observable variables. \begin{itemize} \item Factors may be correlated. \item Errors of reference variables may be correlated. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Student Mental Health Study Design} \framesubtitle{Factors are Anxiety and Depression} \begin{itemize} \item Social variables \item Academic variables \item Psychological assessment variables \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Social Variables} %\framesubtitle{} Using security camera recordings of students eating lunch in the cafeteria (with everyone's permission, of course), the investigators record four social behaviour variables during a designated twenty-minute period. Correlated errors within this set are very likely. \pause \begin{itemize} \item[$d_1$:] Speaking time (not on phone). \item[$d_2$:] Listening time (head turned toward speaker). \item[{\color{blue}$d_5$:}] Number of smiles/laughs while not on cell phone solo\footnote{If two people are looking at a phone together, it's not ``solo," and if they smile or laugh it would be counted}. \item[{\color{blue}$d_6$:}] Time solo on cell phone. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Academic variables} %\framesubtitle{} The following variables are obtained from school records. Measurement errors may not be correlated within this set, but we will be conservative, and assume they might be. In any case, it will be testable. \begin{itemize} \item[$d_3$:] Grade point average last academic session. \item[$d_4$:] Attendance last academic session. \item[{\color{blue}$d_7$:}] Hours per week playing school sports. \item[{\color{blue}$d_8$:}] Hours per week spent on extra-curricular activities, not including school sports. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Psychological Assessment Variables} \framesubtitle{These are the reference variables} \begin{itemize} \item[{\color{red}$d_9$:}] Clinical rating of anxiety. \item[{\color{red}$d_{10}$:}] Clinical rating of depression. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Model for the student mental health example} % Straight from the book %\framesubtitle{} \begin{center} \includegraphics[width=3.5in]{3varOneIndic3} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Moral of the Story} %\framesubtitle{} Divide variables into sets. \begin{itemize} \item Errors may be correlated within sets, but not between sets, as in double measurement. \item But sets need not measure the same things. \pause \item One set consists of reference variables. \item In each set, at least as many variables as factors. \item Arrows from all factors to all variables that are not reference variables. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Bifactor Model from Uli Schimmack's blog} \framesubtitle{No reference variables!} \begin{center} \vspace{-12mm} \includegraphics[width=3in]{Bifactor} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Bifactor Example} % \framesubtitle{} \begin{center} \vspace{-5mm} \includegraphics[width=2.5in]{AnxietyTable} \end{center} {\footnotesize From ``The Rediscovery of Bifactor Measurement Models," by Steven P. Reise, \emph{Multivariate Behavioral Research}, 47:667–696, 2012 --- Used without permission. } % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Second-order Factor Analysis} % Straight from the book %\framesubtitle{} \begin{center} \includegraphics[width=4.5in]{2ndOrder1} \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Copyright Information} This slide show was prepared by \href{http://www.utstat.toronto.edu/~brunner}{Jerry Brunner}, Department of Statistical Sciences, University of Toronto. Except for the table of the Revised Child Anxiety and Depression Scale (which was stolen), it is licensed under a \href{http://creativecommons.org/licenses/by-sa/3.0/deed.en_US} {Creative Commons Attribution - ShareAlike 3.0 Unported License}. Use any other part of it as you like and share the result freely. The \LaTeX~source code is available from the course website: \href{http://www.utstat.toronto.edu/brunner/oldclass/431s23} {\small\texttt{http://www.utstat.toronto.edu/brunner/oldclass/431s23}} \end{frame} \end{document} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % \mathbf{} \boldsymbol{} \boldsymbol{\Sigma}_{} \begin{frame} \frametitle{Frame Title} %\framesubtitle{} \begin{itemize} \item \item \item \end{itemize} \end{frame} {\LARGE \begin{displaymath} \end{displaymath} } % End size % herehere slide %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% Student Mental Health Adding reference variables to an unrestricted factor model p. 306 From The Rediscovery of Bifactor Measurement Models Steven P. Reise, Multivariate Behavioral Research, 47:667–696, 2012 Used without permission.