% Some material is taken from 2015 slide set 15 (Powerpoint), and adapted based on text. % \documentclass[serif]{beamer} % Serif for Computer Modern math font. \documentclass[serif, handout]{beamer} % Handout mode to ignore pause statements \hypersetup{colorlinks,linkcolor=,urlcolor=red} \usefonttheme{serif} % Looks like Computer Modern for non-math text -- nice! \setbeamertemplate{navigation symbols}{} % Suppress navigation symbols % \usetheme{Berlin} % Displays sections on top \usetheme{Frankfurt} % Displays section titles on top: Fairly thin but still swallows some material at bottom of crowded slides %\usetheme{Berkeley} \usepackage[english]{babel} \usepackage{amsmath} % for binom \usepackage{comment} % \usepackage{graphicx} % To include pdf files! % \definecolor{links}{HTML}{2A1B81} % \definecolor{links}{red} \setbeamertemplate{footline}[frame number] \mode \title{Structural Equation Models: The General Case\footnote{See last slide for copyright information.}} \subtitle{STA2053 Fall 2022} \date{} % To suppress date \begin{document} \begin{frame} \titlepage \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Features of Structural Equation Models} %\framesubtitle{} \begin{itemize} \item Multiple equations. \item All the variables are random. \item An explanatory variable in one equation can be the response variable in another equation. \item Models are represented by path diagrams. \item Identifiability is always an issue. \item The statistical models are models of influence. They are often called \emph{causal models}. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{The General (original) Model: {\small Independently for $i=1, \ldots, n$, let}} %\framesubtitle{Also called the ``original" model} \vspace{-8mm} \begin{eqnarray*}\label{original2stage} \mathbf{y}_i &=& \boldsymbol{\alpha} + \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \mathbf{F}_i &=& \left( \begin{array}{c} \mathbf{x}_i \\ \hline \mathbf{y}_i \end{array} \right) \nonumber \\ \mathbf{d}_i &=& \boldsymbol{\nu} + \boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i\pause, \mbox{ where} \end{eqnarray*} \vspace{-2mm} {\scriptsize \begin{itemize} \item $\mathbf{y}_i$ is a $q \times 1$ random vector. \item $\boldsymbol{\alpha}$ is a $q \times 1$ vector of constants. \item $\boldsymbol{\beta}$ is a $q \times q$ matrix of constants \emph{with zeros on the main diagonal}. \item $\boldsymbol{\Gamma}$ is a $q \times p$ matrix of constants. \item $\mathbf{x}_i$ is a $p \times 1$ random vector with expected value $\boldsymbol{\mu}_x$ and positive definite covariance matrix $\boldsymbol{\Phi}_x$. \item $\boldsymbol{\epsilon}_i$ is a $q \times 1$ random vector with expected value zero and positive definite covariance matrix $\boldsymbol{\Psi}$. \item $\mathbf{F}_i$ ($F$ for Factor) is a partitioned vector with $\mathbf{x}_i$ stacked on top of $\mathbf{y}_i$. It is a $(p+q) \times 1$ random vector whose expected value is denoted by $\boldsymbol{\mu}_F$, and whose variance-covariance matrix is denoted by $\boldsymbol{\Phi}$. \item $\mathbf{d}_i$ is a $k \times 1$ random vector. The expected value of $\mathbf{d}_i$ will be denoted by $\boldsymbol{\mu}$, and the covariance matrix of $\mathbf{d}_i$ will be denoted by $\boldsymbol{\Sigma}$. \item $\boldsymbol{\nu}$ is a $k \times 1$ vector of constants. \item $\boldsymbol{\Lambda}$ is a $k \times (p+q)$ matrix of constants. \item $\mathbf{e}_i$ is a $k \times 1$ random vector with expected value zero and covariance matrix $\boldsymbol{\Omega}$, which {\color{red}need not be positive definite}. \item $\mathbf{x}_i$, $\boldsymbol{\epsilon}_i$ and $\mathbf{e}_i$ are independent. \end{itemize} } % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Example: A Path Model with Measurement Error} \begin{columns} \column{0.5\textwidth} \begin{center} \includegraphics[width=2.5in]{mepath} % Ugh! Powerpoint 2015. Sorry! \end{center} \pause \column{0.5\textwidth} %{\footnotesize \begin{eqnarray*} Y_{i,1} &=& \alpha_1 + \gamma_1 X_i + \epsilon_{i,1} \\ Y_{i,2} &=& \alpha_2 + \beta Y_{i,1} + \gamma_2 X_i + \epsilon_{i,2} \\ W_i &=& \nu_1 + \lambda_1X_i + e_{i,1} \\ V_{i,1} &=& \nu_2 + \lambda_2Y{i,1} + e_{i,2} \\ V_{i,2} &=& \nu_3 + \lambda_3Y{i,2} + e_{i,3} \end{eqnarray*} %} % End size \end{columns} \pause {\small \hspace{-3mm}Truth $\approx$ Original Model $\rightarrow$ Surrogate Model 1 $\rightarrow$ Surrogate Model 2 \ldots } % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame}[fragile] \frametitle{Surrogate Mother} %\framesubtitle{} \begin{center} \includegraphics[width=3in]{Harlow.png} {\footnotesize \begin{verbatim} https://en.wikipedia.org/wiki/File:Harlow%27s_Monkey.png \end{verbatim} } % End size \end{center} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Surrogate Models} \pause \framesubtitle{Truth $\approx$ Original Model $\rightarrow$ Surrogate Model 1 $\rightarrow$ Surrogate Model 2 \ldots} \begin{itemize} \item We more or less accept the original model, but we can't identify the parameters. \pause \item So we re-parameterize, obtaining a surrogate model. Repeat. \pause \item We will carefully keep track of the \emph{meaning} of the new parameters in terms of the parameters of the original model. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{The Original Model} %\framesubtitle{} \begin{eqnarray*}\label{original2stage} \mathbf{y}_i &=& \boldsymbol{\alpha} + \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \mathbf{F}_i &=& \left( \begin{array}{c} \mathbf{x}_i \\ \hline \mathbf{y}_i \end{array} \right) \nonumber \\ \mathbf{d}_i &=& \boldsymbol{\nu} + \boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i\pause, \end{eqnarray*} where \ldots \pause \begin{itemize} \item Carefully count the parameters that appear \emph{only} in $E(\mathbf{d}_i)$ and not in $cov(\mathbf{d}_i)$. \pause \item There are more of these parameters than elements of $\mathbf{d}_i$. \item Parameter count rule. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Center the model} %\framesubtitle{} \begin{itemize} \item There are too many expected values and intercepts to identify. \item Center all the random variables in the model by adding and subtracting expected values. \pause \item Obtain a \emph{centered surrogate model} \begin{eqnarray*} \stackrel{c}{\mathbf{y}}_i &=& \boldsymbol{\beta} \! \stackrel{c}{\mathbf{y}}_i + \boldsymbol{\Gamma} \! \stackrel{c}{\mathbf{x}}_i + \boldsymbol{\epsilon}_i \\ \stackrel{c}{\mathbf{F}}_i &=& \left( \begin{array}{c} \stackrel{c}{\mathbf{x}}_i \\ \hline \stackrel{c}{\mathbf{y}}_i \end{array} \right) \\ \stackrel{c}{\mathbf{d}}_i &=& \boldsymbol{\Lambda} \! \stackrel{c}{\mathbf{F}}_i + \mathbf{e}_i \end{eqnarray*} \pause \item Same $\boldsymbol{\beta}$, $\boldsymbol{\Gamma}$ and $\boldsymbol{\Lambda}$, same variances and covariances. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Change of variables} \pause %\framesubtitle{} \begin{itemize} \item Centering is a change of variables. \pause \item Expected values and intercepts are gone, and the dimension of the parameter space is reduced. \pause \item Drop the little $c$ over the random vectors. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{A General Two-Stage Centered Model} \framesubtitle{Stage 1 is the latent variable model and Stage 2 is the measurement model.} Independently for $i=1, \ldots, n$, \begin{eqnarray*} \mathbf{y}_i &=& \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \mathbf{F}_i &=& \left( \begin{array}{c} \mathbf{x}_i \\ \hline \mathbf{y}_i \end{array} \right) \\ \mathbf{d}_i &=& \boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i \end{eqnarray*} \pause \begin{itemize} \item $\mathbf{d}_i$ (the data) are observable. All other variables are latent. \pause \item $\mathbf{y}_i = \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i$ is called the \emph{Latent Variable Model}. \pause \item The latent vectors $\mathbf{x}_i$ and $\mathbf{y}_i$ are collected into a \emph{factor} $\mathbf{F}_i$. \item $\mathbf{d}_i = \boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i$ is called the \emph{Measurement Model}. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{$\mathbf{y}_i= \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i$ ~~ $\mathbf{F}_i = \left( \begin{array}{c} \mathbf{x}_i \\ \hline \mathbf{y}_i \end{array} \right)$ ~~ $\mathbf{d}_i = \boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i$} \pause % \framesubtitle{} \begin{itemize} \item $\mathbf{y}_i$ is a $q \times 1$ random vector. \item $\boldsymbol{\beta}$ is a $q \times q$ matrix of constants with zeros on the main diagonal. \item $\mathbf{x}_i$ is a $p \times 1$ random vector. \item $\boldsymbol{\Gamma}$ is a $q \times p$ matrix of constants. \item $\boldsymbol{\epsilon}_i$ is a $q \times 1$ random vector. \item $\mathbf{F}_i$ ($F$ for Factor) is just $\mathbf{x}_i$ stacked on top of $\mathbf{y}_i$. It is a $(p+q) \times 1$ random vector. \pause \item $\mathbf{d}_i$ is a $k \times 1$ random vector. Sometimes, $\mathbf{d}_i = \left( \begin{array}{c} \mathbf{w}_i \\ \hline \mathbf{v}_i \end{array} \right)$. \pause \item $\boldsymbol{\Lambda}$ is a $k \times (p+q)$ matrix of constants: ``factor loadings." \pause \item $\mathbf{e}_i$ is a $k \times 1$ random vector. \item $\mathbf{x}_i$, $\boldsymbol{\epsilon}_i$ and $\mathbf{e}_i$ are independent. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Covariance matrices} \framesubtitle{More notation} {\small \vspace{-6mm} \begin{eqnarray*} \mathbf{y}_i &=& \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \mathbf{F}_i &=& \left( \begin{array}{c} \mathbf{x}_i \\ \hline \mathbf{y}_i \end{array} \right) \\ \mathbf{d}_i &=& \boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i \\ &&\\ \pause cov(\mathbf{x}_i) &=& \boldsymbol{\Phi}_x \\ cov(\boldsymbol{\epsilon}_i) &=& \boldsymbol{\Psi} \\ cov(\mathbf{F}_i) &=& \boldsymbol{\Phi} = \left( \begin{array}{c c} cov(\mathbf{x}_i) & cov(\mathbf{x}_i,\mathbf{y}_i) \\ cov(\mathbf{y}_i,\mathbf{x}_i) & cov(\mathbf{y}_i) \end{array} \right) = \left( \begin{array}{c c} \boldsymbol{\Phi}_{11} & \boldsymbol{\Phi}_{12} \\ \boldsymbol{\Phi}_{12}^\top & \boldsymbol{\Phi}_{22} \\ \end{array} \right) \\ cov(\mathbf{e}_i) &=& \boldsymbol{\Omega} \\ cov(\mathbf{d}_i) &=& \boldsymbol{\Sigma} \\ \end{eqnarray*} \pause \vspace{-8mm} \begin{itemize} \item Collect the unique elements of $\boldsymbol{\beta}$, $\boldsymbol{\Gamma}$, $\boldsymbol{\Lambda}$, $\boldsymbol{\Phi}_x$, $\boldsymbol{\Psi}$ and $\boldsymbol{\Omega}$ into a parameter vector $\boldsymbol{\theta}$. \item $\boldsymbol{\theta}$ is a \emph{function} of the original model parameters. \end{itemize} } % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} % A huge slide \frametitle{Matrix Form} %\framesubtitle{} %\begin{center} \begin{tabular}{ccc} \includegraphics[width=1in]{mepath} & \raisebox{.45in}{ \begin{minipage}{1.5in} {\footnotesize \begin{eqnarray*} Y_{i,1} &=& \gamma_1 X_i + \epsilon_{i,1} \\ Y_{i,2} &=& \beta Y_{i,1} + \gamma_2 X_i + \epsilon_{i,2} \\ W_i &=& X_i + e_{i,1} \\ V_{i,1} &=& Y_{i,1} + e_{i,2} \\ V_{i,2} &=& Y_{i,2} + e_{i,3} \end{eqnarray*} } % End size \end{minipage} } % End raisebox & \raisebox{.45in}{ \begin{minipage}{1in} {\footnotesize \begin{eqnarray*} \mathbf{y}_i &=& \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \mathbf{F}_i &=& \left( \begin{array}{c} \mathbf{x}_i \\ \hline \mathbf{y}_i \end{array} \right) \\ \mathbf{d}_i &=& \boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i \end{eqnarray*} } % End size \end{minipage} } % End raisebox \end{tabular} \pause \vspace{-3mm} {\small \begin{displaymath} \begin{array}{cccccccccc} \mathbf{y}_i &=& \boldsymbol{\beta} & \mathbf{y}_i & + & \boldsymbol{\Gamma} & \mathbf{x}_i & + & \boldsymbol{\epsilon}_i \\ &&&&&&&&& \\ \left( \begin{array}{c} Y_{i,1} \\ Y_{i,2} \end{array} \right) & = & \left( \begin{array}{c c} 0 & 0 \\ \beta & 0 \\ \end{array} \right) & \left( \begin{array}{c} Y_{i,1} \\ Y_{i,2} \end{array} \right) & + & \left( \begin{array}{c} \gamma_1 \\ \gamma_2 \end{array} \right) & X_i & + & \left( \begin{array}{c} \epsilon_{i,1} \\ \epsilon_{i,2} \end{array} \right) \end{array} \pause \end{displaymath} \begin{displaymath} \begin{array}{cccccc} \mathbf{d}_i &=& \boldsymbol{\Lambda} & \mathbf{F}_i &+& \mathbf{e}_i \\ &&&&& \\ \left( \begin{array}{c} W_i \\ V_{i,1} \\ V_{i,2} \end{array} \right) & = & \left( \begin{array}{c c c} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end{array} \right) & \left( \begin{array}{c} X_i \\ Y_{i,1} \\ Y_{i,2} \end{array} \right) & + & \left( \begin{array}{c} e_{i,1} \\ e_{i,2} \\ e_{i,3} \end{array} \right) \end{array} \end{displaymath} % \pause } % End size % The rest just will not fit. Make it HW. % {\scriptsize % \vspace{2mm} % $cov(\mathbf{X}_i)=Var(X_i)= \phi$ \pause % $cov(\boldsymbol{\epsilon}_i)=\boldsymbol{\Psi} = % \left( \begin{array}{c c} % \psi_1 & 0 \\ % 0 & \psi_2 % \end{array} \right)$ \pause % $cov(\mathbf{e}_i)=\boldsymbol{\Omega} = % \left( \begin{array}{c c c} % \omega_1 & 0 & 0\\ % 0 & \omega_2 & 0 \\ % 0 & 0 & \omega_3 % \end{array} \right)$ %} % End size \end{frame} % End of huge slide %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Observable variables in the ``latent" variable model $\mathbf{y}_i = \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i$ } \framesubtitle{Fairly common} \begin{itemize} \item These present no problem. \item Let $P(e_j=0) = 1$, so $Var(e_j) = 0$. \item And $Cov(e_i,e_j)=0$ \pause \item Because if $P(e_j=0) = 1$, \begin{eqnarray*} Cov(e_i,e_j) &=& E(e_ie_j) - E(e_i)E(e_j) \\ &=& E(e_i\cdot 0) - E(e_i)\cdot 0 \\ &=& 0-0=0 \end{eqnarray*} \pause \item In $\boldsymbol{\Omega} = cov(\mathbf{e}_i)$, column $j$ (and row $j$) are all zeros. \item $\boldsymbol{\Omega}$ singular, no problem. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{What should you be able to do?} %\framesubtitle{} \begin{itemize} \item Given a path diagram, write the model equations and say which exogenous variables are correlated with each other. \item Given the model equations and information about which exogenous variables are correlated with each other, draw the path diagram. \item Given either piece of information, write the model in matrix form and say what all the matrices are. \item Calculate model covariance matrices. \item Check identifiability. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % The next several slides are added to the 2053 lecture. % herehere slide 16 %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{This is a surrogate model} \framesubtitle{For two reasons} \begin{columns} \column{0.5\textwidth} \includegraphics[width=2.5in]{SurPath1} \column{0.5\textwidth} \begin{eqnarray*} Y & = & \beta X + \epsilon \\ W_1 & = & X + e_1 \\ W_2 & = & X + e_2 \\ V & = & Y + e_3 \end{eqnarray*} \end{columns} \begin{columns} \column{0.5\textwidth} \begin{eqnarray*} Var(X) & = & \phi \\ Var(\epsilon) & = & \psi \\ Var(e_1) & = & \omega_1 \\ Var(e_2) & = & \omega_2 \\ Var(e_3) & = & \omega_3 \\ \end{eqnarray*} \pause \column{0.5\textwidth} \vspace{-10mm}%\hspace{10mm}Centered original model \begin{eqnarray*} Y & = & \beta X + \epsilon \\ W_1 & = & \lambda_1 X + e_1 \\ W_2 & = & \lambda_1 X + e_2 \\ V & = & \lambda_2 Y + e_3 \end{eqnarray*} \end{columns} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Covariance Matrices} %\framesubtitle{} \begin{columns} \column{0.5\textwidth} \begin{eqnarray*} Y & = & \beta X + \epsilon \\ W_1 & = & X + e_1 \\ W_2 & = & X + e_2 \\ V & = & Y + e_3 \end{eqnarray*} \column{0.5\textwidth} \hspace{10mm}Centered original model \begin{eqnarray*} Y & = & \beta X + \epsilon \\ W_1 & = & \lambda_1 X + e_1 \\ W_2 & = & \lambda_1 X + e_2 \\ V & = & \lambda_2 Y + e_3 \end{eqnarray*} \end{columns} \begin{columns} \column{0.5\textwidth} {\footnotesize \begin{displaymath} \begin{array}{c|ccc} & W_1 & W_2 & V \\ \hline W_1 & \phi + \omega_1 & \phi & \beta \phi \rule{0mm}{4mm} \\ W_2 & & \phi + \omega_2 & \beta \phi \rule{0mm}{4mm} \\ V & & & \rule{0mm}{4mm} \beta^2\phi + \psi + \omega_3 \\ ~&&& \end{array} \end{displaymath} } % End size \column{0.5\textwidth} {\footnotesize \begin{displaymath} \begin{array}{c|ccc} & W_1 & W_2 & V \\ \hline ~ & \lambda_1^2\phi + \omega_1 & \lambda_1^2\phi & \lambda_1 \lambda_2 \beta \phi \rule{0mm}{4mm}\\ ~ & & \lambda_1^2\phi + \omega_2 & \lambda_1 \lambda_2 \beta \phi \rule{0mm}{4mm}\\ ~ & & & \rule{0mm}{4mm} \lambda_2^2(\beta^2\phi + \psi) \\ ~ & & & ~~~~~ + \omega_3 \end{array} \end{displaymath} } % End size p. 17 \end{columns} \pause {\small \hspace{20mm}Five parameters \hspace{30mm} Seven parameters } \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Change of Variables} %\framesubtitle{} \begin{eqnarray*} W_1 & = & \lambda_1 X + e_1 \\ W_2 & = & \lambda_1 X + e_2 \\ Y & = & \beta X + \epsilon \\ V & = & \lambda_2 Y + e_3 \end{eqnarray*} \pause Let $X^\prime = \lambda_1 X$ \pause \begin{itemize} \item $Var(X^\prime) = \lambda_1^2\phi = \phi^\prime$. \pause \item $W_1 = X^\prime + e_1$ \item $W_2 = X^\prime + e_2$ \end{itemize} \pause \begin{eqnarray*} Y & = & \beta X + \epsilon \\ \pause & = & \left(\frac{\beta}{\lambda_1}\right)(\lambda_1 X) + \epsilon \\ \pause & = & \left(\frac{\beta}{\lambda_1}\right)X^\prime + \epsilon \end{eqnarray*} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Another Change of Variables} \framesubtitle{Have $Y = \left(\frac{\beta}{\lambda_1}\right)X^\prime + \epsilon$} \begin{eqnarray*} V & = & \lambda_2 Y + e_3 \\ \pause & = & \lambda_2 \left(\frac{\beta}{\lambda_1}X^\prime + \epsilon \right)+ e_3 \\ & = & \left(\frac{\lambda_2}{\lambda_1}\beta \right) X^\prime + \lambda_2\epsilon + e_3 \\ \pause & = & \beta^\prime X^\prime + \epsilon^\prime + e_3 \\ \pause & = & Y^\prime + e_3 \end{eqnarray*} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Collect what we have} %\framesubtitle{} \begin{itemize} \item $X^\prime = \lambda_1 X$, $\phi^\prime = \lambda_1^2\phi$. \item $\beta^\prime = \frac{\lambda_2}{\lambda_1}\beta$. \item $\epsilon^\prime = \lambda_2\epsilon$, $\psi^\prime = \lambda_2^2\psi$. \item $\omega_1$, $\omega_2$ and $\omega_3$ are unaffected. \end{itemize} \pause \begin{columns} \column{0.5\textwidth} \begin{eqnarray*} W_1 & = & X + e_1 \\ W_2 & = & X + e_2 \\ Y & = & \beta X + \epsilon \\ V & = & Y + e_3 \end{eqnarray*} \column{0.5\textwidth} \begin{eqnarray*} W_1 & = & X^\prime + e_1 \\ W_2 & = & X^\prime + e_2 \\ Y^\prime & = & \beta^\prime X^\prime + \epsilon^\prime \\ V & = & Y^\prime + e_3 \end{eqnarray*} \end{columns} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Covariance Matrices} %\framesubtitle{} \begin{columns} \column{0.5\textwidth} {\footnotesize \begin{displaymath} \begin{array}{c|ccc} & W_1 & W_2 & V \\ \hline W_1 & \phi + \omega_1 & \phi & \beta \phi \rule{0mm}{4mm} \\ W_2 & & \phi + \omega_2 & \beta \phi \rule{0mm}{4mm} \\ V & & & \rule{0mm}{4mm} \beta^2\phi + \psi + \omega_3 \\ ~&&& \end{array} \end{displaymath} } % End size \column{0.5\textwidth} {\footnotesize \begin{displaymath} \begin{array}{c|ccc} & W_1 & W_2 & V \\ \hline ~ & \lambda_1^2\phi + \omega_1 & \lambda_1^2\phi & \lambda_1 \lambda_2 \beta \phi \rule{0mm}{4mm}\\ ~ & & \lambda_1^2\phi + \omega_2 & \lambda_1 \lambda_2 \beta \phi \rule{0mm}{4mm}\\ ~ & & & \rule{0mm}{4mm} \lambda_2^2(\beta^2\phi + \psi) \\ ~ & & & ~~~~~ + \omega_3 \end{array} \end{displaymath} } % End size \end{columns} \begin{columns} \column{0.5\textwidth} {\footnotesize \begin{displaymath} \begin{array}{c|ccc} & W_1 & W_2 & V \\ \hline W_1 & \phi^\prime + \omega_1 & \phi^\prime & \beta^\prime \phi^\prime \rule{0mm}{4mm} \\ W_2 & & \phi^\prime + \omega_2 & \beta^\prime \phi^\prime \rule{0mm}{4mm} \\ V & & & \rule{0mm}{4mm} \beta^{\prime 2}\phi^\prime + \psi^\prime + \omega_3 \\ ~&&& \end{array} \end{displaymath} } % End size \column{0.5\textwidth} {\scriptsize \begin{displaymath} \begin{array}{c|ccc} & W_1 & W_2 & V \\ \hline ~ & \lambda_1^2\phi + \omega_1 & \lambda_1^2\phi & \left(\frac{\lambda_2}{\lambda_1}\beta \right) (\lambda_1^2\phi) \rule{0mm}{4mm}\\ ~ & & \lambda_1^2\phi + \omega_2 & \left(\frac{\lambda_2}{\lambda_1}\beta \right) (\lambda_1^2\phi) \rule{0mm}{4mm}\\ ~ & & & \rule{0mm}{4mm} \left(\frac{\lambda_2}{\lambda_1}\beta \right)^2(\lambda_1^2\phi) \\ ~ & & & ~~+ (\lambda_2^2\psi) + \omega_3 \end{array} \end{displaymath} } % End size \end{columns} {\small The parameters of the surrogate model are identifiable functions of the original parameters. Mostly.} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{One more re-parameterization} \framesubtitle{By a change of variables} %{\LARGE \begin{eqnarray*} V & = & Y^\prime + e_3 \\ & = & \beta^\prime X^\prime + \epsilon^\prime + e_3 \\ \pause & = & \beta^\prime X^\prime + (\epsilon^\prime + e_3) \\ \pause & = & \beta^\prime X^\prime + \epsilon^{\prime\prime} \end{eqnarray*} \pause with \begin{eqnarray*} Var(\epsilon^{\prime\prime}) & = & Var(\epsilon^\prime) + Var(e_3) \\ & = & \lambda_2^2\psi + \omega_3 \\ & = & \psi^{\prime\prime} \end{eqnarray*} % With $Var(\epsilon^{\prime\prime}) = Var(\epsilon^\prime) + Var(e_3)$ %} % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{The final surrogate model} \framesubtitle{Parameters are identifiable functions of the original model parameters.} \begin{columns} \column{0.5\textwidth} \includegraphics[width=2.2in]{SurPath2} \column{0.5\textwidth} {\small \begin{displaymath} \begin{array}{c|ccc} & W_1 & W_2 & V \\ \hline W_1 & \phi^\prime + \omega_1 & \phi^\prime & \beta^\prime \phi^\prime \\ W_2 & & \phi^\prime + \omega_2 & \beta^\prime \phi^\prime \\ V & & & \beta^{\prime 2}\phi^\prime + \psi^{\prime\prime} \end{array} \end{displaymath} } % End size \end{columns} \pause \vspace{5mm} $\beta^\prime = \frac{\lambda_2}{\lambda_1}\beta$ \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Next slide please} %\framesubtitle{} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Recall the notation} \begin{eqnarray*} \mathbf{y}_i &=& \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \mathbf{F}_i &=& \left( \begin{array}{c} \mathbf{x}_i \\ \hline \mathbf{y}_i \end{array} \right) \\ \mathbf{d}_i &=& \boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i \\ &&\\ cov(\mathbf{x}_i) &=& \boldsymbol{\Phi}_x \\ cov(\boldsymbol{\epsilon}_i) &=& \boldsymbol{\Psi} \\ cov(\mathbf{F}_i) &=& \boldsymbol{\Phi} = \left( \begin{array}{c c} cov(\mathbf{x}_i) & cov(\mathbf{x}_i,\mathbf{y}_i) \\ cov(\mathbf{y}_i,\mathbf{x}_i) & cov(\mathbf{y}_i) \end{array} \right) = \left( \begin{array}{c c} \boldsymbol{\Phi}_{11} & \boldsymbol{\Phi}_{12} \\ \boldsymbol{\Phi}_{12}^\top & \boldsymbol{\Phi}_{22} \\ \end{array} \right) \\ cov(\mathbf{e}_i) &=& \boldsymbol{\Omega} \\ cov(\mathbf{D}_i) &=& \boldsymbol{\Sigma} \\ \end{eqnarray*} \pause Calculate a general expression for $\boldsymbol{\Sigma}(\boldsymbol{\theta})$. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{For the latent variable model, calculate $\boldsymbol{\Phi} = cov(\mathbf{F}_i)$} \framesubtitle{Have $cov(\mathbf{x}_i) = \boldsymbol{\Phi}_x$, need $cov(\mathbf{y}_i)$ and $cov(\mathbf{x}_i,\mathbf{y}_i)$} \pause %{\footnotesize \begin{eqnarray*} & & \mathbf{y}_i = \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \pause &\Rightarrow& \mathbf{y}_i - \boldsymbol{\beta} \mathbf{y}_i = \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \pause &\Rightarrow& \mathbf{Iy}_i - \boldsymbol{\beta} \mathbf{y}_i = \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \pause &\Rightarrow& (\mathbf{I} - \boldsymbol{\beta} )\mathbf{y}_i = \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i \\ \pause & {\color{red} \Rightarrow} & (\mathbf{I} - \boldsymbol{\beta} )^{-1}(\mathbf{I} - \boldsymbol{\beta} )\mathbf{y}_i = (\mathbf{I} - \boldsymbol{\beta} )^{-1}(\boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i) \\ \pause &\Rightarrow& \mathbf{y}_i = (\mathbf{I} - \boldsymbol{\beta} )^{-1}(\boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i) \end{eqnarray*} \pause So, \begin{eqnarray*} cov(\mathbf{y}_i) \pause &=& (\mathbf{I} - \boldsymbol{\beta} )^{-1}cov(\boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i)(\mathbf{I} - \boldsymbol{\beta} )^{-1\top} \\ \pause &=& (\mathbf{I} - \boldsymbol{\beta} )^{-1}\left(cov(\boldsymbol{\Gamma} \mathbf{x}_i) + cov(\boldsymbol{\epsilon}_i)\right) (\mathbf{I} - \boldsymbol{\beta}^\top )^{-1} \\ \pause &=& (\mathbf{I} - \boldsymbol{\beta} )^{-1} \left(\boldsymbol{\Gamma}\boldsymbol{\Phi}_x\boldsymbol{\Gamma}^\top + \boldsymbol{\Psi} \right) (\mathbf{I} - \boldsymbol{\beta}^\top )^{-1} \end{eqnarray*} %} % End size \end{frame} % That's a lot cleaner than the way I was doing it before. %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Theorem: If the original model holds, $(\mathbf{I} - \boldsymbol{\beta} )^{-1}$ exists.} %\framesubtitle{} $\mathbf{y}_i = \boldsymbol{\alpha} + \boldsymbol{\beta} \mathbf{y}_i + \boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i$ yields $ (\mathbf{I} - \boldsymbol{\beta} )\mathbf{y}_i = \boldsymbol{\alpha} +\boldsymbol{\Gamma} \mathbf{x}_i + \boldsymbol{\epsilon}_i$. \pause Suppose $(\mathbf{I} - \boldsymbol{\beta} )^{-1}$ does not exist. \pause Then the rows of $\mathbf{I} - \boldsymbol{\beta}$ are linearly dependent, and there is a $q \times 1$ non-zero vector of constants $\mathbf{a}$ with $\mathbf{a}^\top (\mathbf{I} - \boldsymbol{\beta} ) = 0$. \pause So, \begin{eqnarray*} \mathbf{a}^\top(\mathbf{I} - \boldsymbol{\beta} )\mathbf{y}_i &=& 0 \pause = \mathbf{a}^\top\boldsymbol{\alpha} + \mathbf{a}^\top\boldsymbol{\Gamma} \mathbf{x}_i + \mathbf{a}^\top\boldsymbol{\epsilon}_i \\ \pause \Rightarrow Var(0) &=& Var(\mathbf{a}^\top\boldsymbol{\Gamma} \mathbf{x}_i) + Var(\mathbf{a}^\top\boldsymbol{\epsilon}_i) \\ \pause \Rightarrow 0 &=& \mathbf{a}^\top \boldsymbol{\Gamma \Phi}_x \boldsymbol{\Gamma}^\top \mathbf{a} + \mathbf{a}^\top \boldsymbol{\Psi}\mathbf{a} \pause > 0. \end{eqnarray*} \pause Contradicts $\mathbf{I} - \boldsymbol{\beta}$ singular. %But the quantity on the right side is strictly positive, because while $\boldsymbol{\Gamma \Phi}_x \boldsymbol{\Gamma}^\top = cov(\boldsymbol{\Gamma} \mathbf{x}_i)$ is only guaranteed to be non-negative definite, $\boldsymbol{\Psi}$ is strictly positive definite according to the model. Thus, the assumption that $\mathbf{I} - \boldsymbol{\beta}$ is singular leads to a contradiction. This shows that $(\mathbf{I} - \boldsymbol{\beta} )^{-1}$ must exist if the model holds. ~$\blacksquare$ \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{A hole in the parameter space} %\framesubtitle{} $|\mathbf{I} - \boldsymbol{\beta}| \neq 0$ can create a hole in the parameter space. \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{More calculations} %\framesubtitle{} \begin{itemize} \item Have $ cov(\mathbf{y}_i) = (\mathbf{I} - \boldsymbol{\beta} )^{-1} \left(\boldsymbol{\Gamma}\boldsymbol{\Phi}_x\boldsymbol{\Gamma}^\top + \boldsymbol{\Psi} \right) (\mathbf{I} - \boldsymbol{\beta}^\top )^{-1}$. \pause \item Know $cov(\mathbf{x}_i) = \boldsymbol{\Phi}_x$ \item Easy to get $cov(\mathbf{x}_i,\mathbf{y}_i)$. \pause \item[] \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{For the measurement model, calculate $\boldsymbol{\Sigma} = cov(\mathbf{d}_i)$} %\framesubtitle{} %{\LARGE \begin{eqnarray*} \mathbf{d}_i &=& \boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i \\ \pause \Rightarrow cov(\mathbf{d}_i) &=& cov(\boldsymbol{\Lambda}\mathbf{F}_i + \mathbf{e}_i) \\ \pause &=& cov(\boldsymbol{\Lambda}\mathbf{F}_i) + cov(\mathbf{e}_i) \\ \pause &=& \boldsymbol{\Lambda}cov(\mathbf{F}_i)\boldsymbol{\Lambda}^\top + cov(\mathbf{e}_i) \\ \pause &=& \boldsymbol{\Lambda}\boldsymbol{\Phi}\boldsymbol{\Lambda}^\top + \boldsymbol{\Omega} \\ \pause &=& \boldsymbol{\Sigma} \end{eqnarray*} %} % End size \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Two-stage Proofs of Identifiability} \framesubtitle{Stage 1 is the latent variable model and Stage 2 is the measurement model.} \pause \begin{itemize} \item Show the parameters of the latent variable model $(\boldsymbol{\beta}, \boldsymbol{\Gamma}, \boldsymbol{\Phi}_x, \boldsymbol{\Psi})$ can be recovered from $\boldsymbol{\Phi} = cov(\mathbf{F}_i)$. \pause \item Solve $\left( \begin{array}{c c} cov(\mathbf{x}_i) & cov(\mathbf{x}_i,\mathbf{y}_i) \\ cov(\mathbf{y}_i,\mathbf{x}_i) & cov(\mathbf{y}_i) \end{array} \right) = \boldsymbol{\Phi} = \left( \begin{array}{c c} \boldsymbol{\Phi}_{11} & \boldsymbol{\Phi}_{12} \\ \boldsymbol{\Phi}_{12}^\top & \boldsymbol{\Phi}_{22} \\ \end{array} \right)$ for $(\boldsymbol{\beta}, \boldsymbol{\Gamma}, \boldsymbol{\Phi}_x, \boldsymbol{\Psi})$? \pause \item Show the parameters of the measurement model $(\boldsymbol{\Lambda},\boldsymbol{\Phi},\boldsymbol{\Omega})$ can be recovered from $\boldsymbol{\Sigma} = cov(\mathbf{d}_i)$. \pause \item This means all the parameters can be recovered from $\boldsymbol{\Sigma}$. \pause \item Break a big problem into two smaller ones. \item Develop \emph{rules} for checking identifiability at each stage. \item Just look at the path diagram. \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{frame} \frametitle{Copyright Information} This slide show was prepared by \href{http://www.utstat.toronto.edu/~brunner}{Jerry Brunner}, Department of Statistical Sciences, University of Toronto. It is licensed under a \href{http://creativecommons.org/licenses/by-sa/3.0/deed.en_US} {Creative Commons Attribution - ShareAlike 3.0 Unported License}. Use any part of it as you like and share the result freely. The \LaTeX~source code is available from the course website: \vspace{3mm} \href{http://www.utstat.toronto.edu/brunner/oldclass/2053f22} {\small\texttt{http://www.utstat.toronto.edu/brunner/oldclass/2053f22}} \end{frame} \end{document} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% {\LARGE \begin{displaymath} \end{displaymath} } \begin{frame} \frametitle{} %\framesubtitle{} \begin{itemize} \item \item \item \end{itemize} \end{frame} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%