\documentclass[10pt]{report} \usepackage{amsbsy} % for \boldsymbol and \pmb \usepackage{amssymb} % for \blacksquare \usepackage{amsmath} \usepackage[colorlinks=true, pdfstartview=FitV, linkcolor=blue, citecolor=blue, urlcolor=blue]{hyperref} % For links \oddsidemargin=0in % Good for US Letter paper \evensidemargin=0in \textwidth=6.3in \topmargin=-1.0in % Was 0.5 \headheight=0.2in \headsep=0.5in \textheight=8.5in \pagestyle{empty} % No page numbers \begin{document} \begin{center} {\textbf{Parameter Identifiability Rules}}\footnote{This handout was prepared by \href{http://www.utstat.toronto.edu/~brunner}{Jerry Brunner}, Department of Statistical Sciences, University of Toronto. It is licensed under a \href{http://creativecommons.org/licenses/by-sa/3.0/deed.en_US} {Creative Commons Attribution - ShareAlike 3.0 Unported License}. Use any part of it as you like and share the result freely. The \LaTeX~source code is available from the course website: \href{http://www.utstat.toronto.edu/brunner/oldclass/431s23} {\small\texttt{http://www.utstat.toronto.edu/brunner/oldclass/431s23}}} \\ % These would be better with some consolidation of standardized and unstandardized. \end{center} \vspace{2mm} \noindent \textbf{Note:} All the rules listed here assume that errors are independent of exogenous variables that are not errors, and that the variables have been centered to have expected value zero. \begin{enumerate} \item \textbf{Parameter Count Rule:} If a model has more parameters than covariance structure equations, the parameter vector can be identifiable on at most a set of volume zero in the parameter space. This applies to all models. \item \emph{Latent variable model}: $\mathbf{y}_i = \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i$~ Here, identifiability means that the parameters involved are functions of $cov(\mathbf{F}_i)=\boldsymbol{\Phi}$. \begin{enumerate} \item \textbf{Regression Rule:} If no endogenous variables influence other endogenous variables, the model parameters are identifiable. \item \textbf{Acyclic Rule}: Parameters of the Latent Variable Model are identifiable if the model is acyclic (no feedback loops through straight arrows) and the following conditions hold. \begin{itemize} \item Organize the variables that are not error terms into sets. Set 0 consists of all the exogenous variables. \item For $j=1,\ldots ,k$, each endogenous variable in set $j$ is influenced by at least one variable in set $j-1$, and also possibly by variables in earlier sets. \item Error terms for the variables in a set may have non-zero covariances. All other covariances between error terms are zero. These conditions are satisfied if $\boldsymbol{\Psi}$ is diagonal. \end{itemize} \end{enumerate} % End of latent variable model rules \item \emph{Measurement model} (Factor analysis) In these rules, latent variables that are not error terms are described as ``factors." \noindent \textbf{Definition}: A \emph{reference variable} for a latent variable is an observable variable that is a function only of that latent variable and an error term. The factor loading is non-zero. \begin{enumerate} \item \textbf{Double Measurement Rule}: Parameters of the double measurement model are identifiable. All factor loadings equal one. Correlated measurement errors are allowed within sets of measurements, but not between sets. \item \textbf{Three-Variable Rule}: For a factor analysis model with a single factor, the parameters will be identifiable provided that \begin{itemize} \item There are at least three reference variables. \item Either the factor is standardized and the sign one factor loading is known, or else at least one factor loading equals one. \item Errors are independent of one another and of the factors. \end{itemize} \item \textbf{Four-variable 2-factor Rule}: The parameters of a factor analysis model are identifiable provided \begin{itemize} \item There are four observable variables and two factors. \item There are two reference variables for each factor. \item For each factor, either the variance equals one and the sign of one factor loading is known, or the factor loading of at least one reference variable is equal to one. \item The two factors have non-zero covariance. \item Errors are independent of one another and of the factors. \end{itemize} \newpage %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \enlargethispage*{1000 pt} \item \textbf{Two-variable Addition Rule}: A factor with just two reference variables may be added to a measurement model whose parameters are identifiable, and the parameters of the combined model will be identifiable provided \begin{itemize} \item The errors for the two additional variables are independent of one another and of the error terms already in the model. \item Either the variance of the additional factor equals one and the sign of one factor loading is known, or the factor loading of at least one reference variable is equal to one. \item In the existing model with identifiable parameters, \begin{itemize} \item There is at least one reference variable for each factor, and \item At least one factor has a non-zero covariance with the new factor. \end{itemize} \end{itemize} \item \textbf{Vector 3-variable Rule}: % Versions of this were called Joreskog's rule and then the reference variables rule. Let $\mathbf{d}_1 = \mathbf{F} + \mathbf{e}_1$, $\mathbf{d}_2 = \boldsymbol{\Lambda}_2\mathbf{F} + \mathbf{e}_2$ and $\mathbf{d}_3 = \boldsymbol{\Lambda}_3\mathbf{F} + \mathbf{e}_3$, where $\mathbf{F}$, $\mathbf{d}_1$ and $\mathbf{d}_2$ and $\mathbf{d}_3$ are all $p \times 1$, $\boldsymbol{\Lambda}_2$ and $\boldsymbol{\Lambda}_3$ have inverses, and $\mathbf{F}$ and the $\mathbf{e}_j$ are all independent. Then all the parameters are identifiable. \ \item \textbf{Factor Model Combination Rule}: % See assignment 10, Q13 Suppose there are two factor analysis models $A$ and $B$ with identifiable parameters. Parameters of the combined model will be identifiable provided \begin{itemize} \item Every factor has at least one reference variable. \item The error terms of the reference variables in set $A$ have zero covariance with the error terms of the reference variables in set in $B$. \end{itemize} The combined model includes covariances between factors, and also potentially covariances between error terms from the two sets, except as noted above. \item \textbf{Extra Variables Rule}: A set of observable variables may be added to a measurement model whose parameters are \emph{already identifiable}, provided \begin{itemize} \item There is a reference variable for each factor in the existing model. \item Error terms of the additional variables have zero covariance with the error terms of the reference variables in the existing model. % \item Error terms of the additional variables have zero covariance with the factors. \end{itemize} Under these conditions, \begin{itemize} \item Straight arrows with factor loadings on them may point from each existing factor to each new variable. \item Error terms for the new set of variables may have non-zero covariances with each other, and with the error terms in the original model that do not belong to the reference variables. \item You don't need to include all such links. \end{itemize} \item \textbf{Equivalence Rule}: For a factor analysis model with at least one reference variable for each factor, suppose that surrogate models are obtained by either standardizing the factors, or by setting the factor loading of a reference variable equal to one for each factor. Then the parameters of one surrogate model are identifiable if and only if the parameters of the other surrogate model are identifiable. \item \textbf{Error-Free Rule}: A vector of observable variables may be added to the factors of a measurement model whose parameters are identifiable, provided that the new observable variables are independent of the errors in the measurement model, and there is at least one reference variable for each factor in the existing model. Parameters of the combined moel are identifiable. % The practical consequence is that variables assumed to be measured without error may be included in the latent component of a structural equation model, provided that the measurement model for the other variables has identifiable parameters. \end{enumerate} % End of measurement model rules \item \textbf{Two-Step Rule}: This applies to models with both a measurement component and a latent variable component, including the full two-stage structural equation model. \begin{itemize} \item[1:] Consider the latent variable model as a model for observed variables. Check identifiability (usually using the Regression Rule and the Acyclic Rule). \item[2:] Consider the measurement model as a factor analysis model, ignoring the structure of $cov(\mathbf{F}_i)$. Check identifiability. \end{itemize} If both identification checks are successful, the parameters of the combined model are identifiable. \end{enumerate} % End of ALL the rules. \end{document}