Skip to Main Content

ABSTRACT

We discuss the Gaussian graphical model (GGM; an undirected network of partial correlation coefficients) and detail its utility as an exploratory data analysis tool. The GGM shows which variables predict one-another, allows for sparse modeling of covariance structures, and may highlight potential causal relationships between observed variables. We describe the utility in three kinds of psychological data sets: data sets in which consecutive cases are assumed independent (e.g., cross-sectional data), temporally ordered data sets (e.g., n = 1 time series), and a mixture of the 2 (e.g., n > 1 time series). In time-series analysis, the GGM can be used to model the residual structure of a vector-autoregression analysis (VAR), also termed graphical VAR. Two network models can then be obtained: a temporal network and a contemporaneous network. When analyzing data from multiple subjects, a GGM can also be formed on the covariance structure of stationary means—the between-subjects network. We discuss the interpretation of these models and propose estimation methods to obtain these networks, which we implement in the R packages graphicalVAR and mlVAR. The methods are showcased in two empirical examples, and simulation studies on these methods are included in the supplementary materials.

There has been a surge of network models being applied to psychological data sets in recent years. This is consistent with a general call to conceptualize observed psychological processes not merely as indicative of latent common causes but rather as emergent behavior of complex, dynamical systems in which psychological, biological, and sociological components directly affect each other (Borsboom, Cramer, Schmittmann, Epskamp, & Waldorp, 2011; Cramer et al., 2012; Cramer, Waldorp, van der Maas, & Borsboom, 2010; Schmittmann et al., 2013; van der Maas et al., 2006). Such relationships are typically not known, and probabilistic network models (Koller & Friedman, 2009) are used to explore potential dynamical relationships between observables (Epskamp, Maris, Waldorp, & Borsboom, in press; van Borkulo et al., 2014). In this paper, we aim to provide a methodological introduction to a powerful probabilistic network model applicable in exploratory data analysis, the Gaussian graphical model (GGM), and to propose how it can be used and interpreted in the analysis of both cross-sectional and time-series data.

Two lines of network research in psychology. We can currently distinguish two distinct and mostly separate lines of research in which networks are utilized on psychological data sets: the modeling of cross-sectional data and the modeling of intensive repeated measures in relatively short time frames (e.g., several times per day during several weeks). In cross-sectional modeling, a model is applied to a data set in which multiple subjects are measured only once. The most popularly used methods estimate undirected network models—so-called pairwise Markov random fields (Epskamp et al., in press; Murphy, 2012). When the data are continuous and assumed normally distributed, the GGM can be estimated. The GGM estimates a network of partial correlation coefficients—the correlation between two variables after conditioning on all other variables in the data set (Epskamp, Borsboom, & Fried, 2017a). This model is applied extensively to psychological data (e.g., Cramer et al., 2012; Fried, Epskamp, Nesse, Tuerlinckx, & Borsboom, 2016; Isvoranu et al., 2017; Kossakowski et al., 2015; McNally et al., 2015; van Borkulo et al., 2015).

Researchers can obtain time-series data by using the experience sampling method (ESM; Myin-Germeys et al., 2009), in which subjects are asked several times per day to fill out a short questionnaire using a device or smartphone app. Also, time-series data can arise from diary studies (e.g., a questionnaire completed at the end of the day) or physiological measurements, among other methods. Often, repeated measures of one or multiple participants are modeled through the use of (multilevel) vector autoregressive (VAR) models, which estimate how well each variable predicts the measured variables at the next time point (Borsboom & Cramer, 2013). These models are increasingly popular in assessing intraindividual dynamical structures (e.g., Bringmann et al., 2013; Bringmann, Lemmens, Huibers, Borsboom, & Tuerlinckx, 2015; Wigman et al., 2015).

Estimating the GGM is not limited to cross-sectional data; the model merely does not take temporal information into account. As such, the lines of research on network modeling of cross-sectional data and time-series data can naturally be combined. First, GGM models can readily be estimated on repeated measures, if these can be assumed to be temporally independent. Second, as the VAR model can be seen as a generalization of the GGM that takes violations of independence between consecutive cases into account; the GGM can be used to model the residual (innovation) structure of a VAR model to gain insight in the contemporaneous time level of a time-series analysis. Finally, the between-subjects effects of n > 1 studies can also be modeled through the use of the GGM.

Outline.We show that in time-series modeling the GGM allows researchers to extend the modeling framework to incorporate contemporaneous and between-subjects effects. We do this by building up the model complexity in three steps: (1) when cases can be assumed to be independent (e.g., cross-sectional data or repeated measures in which no auto-regression is assumed), (2) temporally ordered data (e.g., n = 1 time-series data or n > 1 time-series data where no individual differences are assumed), and (3) temporally ordered data from multiple subjects (e.g., n > 1 time series). The final level of model complexity leads to a novel contribution of this paper: separation of variance into contemporaneous, temporal, and between-subjects network structures. We propose novel estimation procedures to estimate these models, which we have implemented in two free software packages: mlVAR,1 and graphicalVAR.2 We furthermore expand on existing literature by providing a comprehensive methodological discussion of the GGM, by comparing the GGM to structural equation modeling (SEM; Kaplan, 2000; Wright, 1921), by providing overviews of estimation methods and software packages useable in each kind of data set and by discussing the interpretation of networks estimated at the contemporaneous and between-subjects levels. We showcase network models estimated from n > 1 time-series data in two empirical examples by reanalyzing existing data sets (Bringmann et al., 2013; Geschwind et al., 2011; Mõttus, Epskamp, & Francis, 2017). In the supplementary materials, we provide codes to perform the analyses and we assess the performance of these methods in large-scale simulation studies. To aid the reader in the various different terms used in this paper, we have included a glossary of terms in the Appendix.

1. The Gaussian graphical model

Let denote a random vector with as its realization.3 We assume is centered4 and normally distributed with some variance-covariance matrix : (1) The subscript C denotes a case (a row in the spreadsheet). We currently do not define the nature of the observed variables. Thus, can consist of variables relating to one or more subjects, could contain repeated measures on one or more variables, could contain variables of a single subject that do not vary within-subject, and so forth. Consider three examples: (1) Y1 could represent the level of anxiety of subject p on day 1 and Y2 the level of anxiety of subject p on day 2, (2) Y1 could represent the length of subject p and Y2 the number of times subject p bumps his or her head, and (3) Y1 could represent the number of cigarettes subject p smokes per day and Y2 the number of cigarettes another subject p + 1 smokes per day (case C then represents a dyadic pair).

Partial correlation networks. Assuming multivariate normality, encodes all the information necessary to determine how the observed measures relate to one another. However, we will not focus on in this paper but rather on its inverse—the precision matrix : Of particular importance is that the precision matrix can be standardized to encode partial correlation coefficients of two variables, given all other variables (dropping subscript C for notational clarity; Lauritzen 1996):5 (2) in which κij denotes an element of , and denotes the set of variables without i and j. These partial correlations can be graphically displayed in a weighted network, in which each variable Yi is represented as a node, and connections (edges) between these nodes represent the partial correlation between two variables. When the partial correlation (thus the corresponding element in ) equals zero, no edge is drawn. Thus, modeling the inverse variance-covariance matrix, such that every nonzero element is treated as a freely estimated parameter, allows for a sparse model for (i.e., every element in may be nonzero while some elements in are zero; Epskamp, Rhemtulla, & Borsboom 2017d). Such a model is termed a GGM (Lauritzen, 1996). Of note, when the sample-variance-covariance matrix is inverted and standardized, no partial correlation will be exactly equal to zero and the GGM will therefore be saturated. To obtain a sparse model with testable implications, in this paper partial correlations are forced to zero either by using thresholding rules or regularization techniques.

When drawing a GGM as a network (often termed a partial correlation network), positive partial correlations are typically visualized with blue or green edges and negative partial correlations with red edges,6 and the absolute strength of a partial correlation is represented by the width and saturation of an edge (Epskamp et al., 2012). When a partial correlation is zero, we draw no edge between two nodes. As such, the GGM can be seen as a network model of conditional associations; no edge indicates that two variables are independent after conditioning on all other variables in the data set. This allows us to model conditional associations, which we might expect to be zero, rather than marginal associations, which we rarely expect to be zero (Meehl, 1990).

To exemplify the above, suppose for three variables “fatigue,” “concentration problems,” and “insomnia,” the true variance-covariance matrix is To model this matrix, we need six parameters (three covariances and three variances). The corresponding true precision matrix becomes Similar to SEM, a model can be devised that perfectly explains this pattern using only five parameters, because one of the elements in can be constrained to be zero (Epskamp et al., 2017d). We can now standardize this matrix and make the off-diagonal elements negative (Equation (2)) to obtain the partial correlation matrix, which we will denote : This matrix can be used to draw a network as is shown in Figure 1. This figure shows that someone who is tired is also more likely to suffer from concentration problems and insomnia. Furthermore, this network shows that the correlation between insomnia and concentration problems can be explained by the relationships of both variables with fatigue: concentration problems and insomnia are conditionally independent given the level of fatigue.

Figure 1. A hypothetical example of a GGM on psychological variables. Nodes represent someone’s ability to concentrate, someone’s level of fatigue, and someone’s level of insomnia. Connections between the nodes, termed edges, represent partial correlation coefficients between two variables after conditioning on the third. Blue edges indicate positive partial correlations, red edges indicate negative partial correlations, and the width and saturation of an edge corresponds to the absolute value of the partial correlation.

Interpreting GGMs. This paper concerns the exploratory estimation of GGMs from various sources of data, without prior knowledge on the model structure. Such undirected network models can be interpreted in strikingly different ways, ranging from a no causal interpretation to a strong causal interpretation:

(1)

Predictive effects. The GGM can be interpreted without any causal interpretation and used merely as a tool to show which variables predict one-another. Interpreting the parameters associated with the model ABC requires a causal interpretation, while the predictive quality between these nodes can directly be obtained from the equivalent GGM ABC: only information on node B is needed when predicting A or C. As such, the GGM can always be interpreted to show predictive effects and offers a powerful exploratory tool to map out multicollinearity.

(2)

Indicative of causal effects. The GGM is closely tied to causal modeling. If a causal model between observed variables generated the data, then an edge AB appears in the GGM only if there is a causal link between the variables (e.g., AB or AB), or if both variables cause a third variable in the data (e.g., ACB). Exploratory estimation of such models relies on stringent assumptions (e.g., acyclicity), suffers from a problem of many equivalent models, and may lead to over-saturated models. The GGM, on the other hand, is well identified and does not feature equivalent models. Therefore, at the cost of losing information on the direction of effect, exploratory search algorithms perform well in identifying a GGM. Because of this close tie to causal modeling, edges in the GGM may be interpreted as indicative of potential causal pathways.

(3)

Causal generating model. Undirected network models have a long history of being used as data-generating models in diverse scientific fields such as statistical physics (Murphy, 2012). For example, in a simple ferromagnetic Ising model of two particles that tend to be aligned (Epskamp et al., in press), AB, intervening on A would impact B and intervening on B would impact A. To this end, undirected network models allow for a unique causal interpretation: one of genuine symmetric effects. This interpretation is discussed often in the literature on network psychometrics, and used in complexity research demonstrating emergent phenomena (e.g., the positive manifold or phase transitions) that may occur in such a network of cellular automata (Cramer et al., 2016; Dalege et al., 2016; Kruis & Maris, 2016; van der Maas et al., 2006).

In addition, the GGM is closely tied to factor analysis, allowing for extensions to factor modeling through the use of network modeling (Epskamp et al., 2017d). The main focus of this paper is discussing the second interpretation, while also describing how the GGM may be used to show predictive effects. We detail these first two points below by first showing how partial correlation coefficients correspond to multiple regression coefficients and next discussing the relationships between the GGM and SEM. Point 3 follows from observing that the GGM is directly related to similar undirected models such as the Ising model (Ising, 1925). A discussion on the causal interpretation of such models is beyond the scope of this paper, and we refer the reader for this topic to Epskamp et al. (in press) and van Borkulo et al. (2014).

1.1. The Gaussian graphical model and multiple regressions

An edge in a GGM indicates that one node predicts a connected node after controlling for all other nodes in the network. This can also be shown in the relationship between coefficients obtained from least-squares prediction and the inverse variance-covariance matrix. Let represent an k × k matrix with zeros on the diagonal. Furthermore, let represent the ith row of without the ith element (as the diagonal is set to zero), which contains the regression coefficients obtained in a multiple regression model: As such, γij encodes how well the jth variable predicts the ith variable. This predictive effect is naturally symmetric; if knowing someone’s level of insomnia predicts his or her level of fatigue, then conversely knowing someone’s level of fatigue allows us to predict his or her level of insomnia. As a result, γij is proportional to γji. There is a direct relationship between these regression coefficients and the inverse variance-covariance matrix (Meinshausen & Bühlmann, 2006). Let denote a diagonal matrix on which the ith diagonal element is the inverse of the ith residual variance: dii = 1/Var(ϵCi). It can then be shown (Pourahmadi, 2011) that7 (3) Thus, κij is proportional to both γij and γji; a zero in the inverse variance-covariance matrix indicates that one variable does not predict another. Consequently, the network tells us something about the extent to which variables predict each other. This predictive quality is the cornerstone for how such network models are often applied (Hastie, Tibshirani, & Wainwright, 2015), for example, in recommender-systems that recommend users on products they might like depending on which products the user already liked (Marsman, Waldorp, & Maris, 2017b). In addition to these applications and aiding the interpretation of GGM models, this relationship between multiple regression and undirected network edges plays a crucial role in many network estimation procedures (Haslbeck & Waldorp, 2016b; Meinshausen & Bühlmann, 2006; van Borkulo et al., 2014), including the methods discussed below in this paper.

1.2. The Gaussian graphical model and structural equation modeling

Let represent a set of unobserved variables, which we assume to be jointly normally distributed with . Then, we can form an encompassing framework for several possible generating models:8 (4) in which is a diagonal matrix,9 indicating that after conditioning on all causes the variables are independent, is a square matrix with zeros on the diagonal of causal effects between observed variables, and is a factor-loading matrix. The variance-covariance matrix of may in turn be modeled in various ways to achieve complicated model setups. The expression above is well-known in SEM, which allows for confirmatory testing of causal models. In exploratory estimation, one could assume no latent variables exist and aim estimate (causal models), or one could assume no relationships between observed variables exist and aim to estimate (factor models). We contrast both to the GGM below.

1.2.1. Causal models

Suppose there are no unobserved causes to any of the variables in , and the variables in are only caused by other variables in . The corresponding model for becomes: (5) In this expression, can now be seen to encode the causal model (Pearl, 2000). Table 1 summarizes the comparison between such causal models and GGMs. Although useful for generating data and confirmatory testing, we can see two problems in exploratory estimation of without any prior knowledge. First, if m variables are included, contains m(m + 1)/2 elements, while contains m parameters and contains m(m − 1) parameters. As a result, the model above is underidentified without stringent restrictions on . One assumption is that can be ordered such that is lower triangular, indicating that if this matrix is used to draw a directed graph—a graph in which AB indicates that A causes B—that graph does not contain any cycles, meaning that directed edges cannot be traced from any node back to itself (e.g., ABA). Such a graph is called a directed acyclic graph (DAG; Kalisch & Bühlmann, 2007; Pearl, 2000). If repeated measures are available at the correct time scale, reciprocal effects and cycles can often be adequately modeled as acyclic effects unfolding over time. Without such information, cycles can be modeled and can be identified when exogenous variables are present (such as the weather, time, or, depending on the modeling framework, lagged variables; Rigdon, 1995), but the interpretation of such cycles is not without problems (Hayduk, 2009). Several software packages exist that aim to find such a DAG (e.g., pcalg, Kalisch, Mächler, Colombo, Maathuis, & Bühlmann, 2012; bnlearn, Scutari, 2010). The assumption of acyclicity, however, is debatable in the context of psychological variables (Schmittmann et al., 2013) because many effects can be plausibly assumed cyclic (e.g., fatigue → concentration problems → stress → fatigue).

Table 1. Overview of causal models (directed networks) and Gaussian graphical models (undirected networks).

Second, the same structure for can be obtained under various different specifications of . Thus, many equivalent models can lead to exactly the same fit. This can be seen because several matrix decompositions of , such as a Cholesky decomposition or an eigendecomposition, can be used to produce equivalently fitting . The problem of equivalent models is also well-known in the literature on directed networks and SEM (MacCallum, Wegener, Uchino, & Fabrigar, 1993; Pearl, 2000). For example, the following three causal models are not statistically distinguishable:

(1)

Concentration → Fatigue → Insomnia

(2)

Concentration ← Fatigue → Insomnia

(3)

Concentration ← Fatigue ← Insomnia

All three models only imply that concentration and insomnia are conditionally independent given fatigue. With more variables, the number of potential equivalent models increases drastically, making it evident that model search is likely to fail. At best, exploratory estimation can result in a set of equally plausible DAGs (an equivalence class; Drton & Maathuis, 2017), each differently parameterized and each leading to different strong causal hypotheses.

Causal modeling and the GGM. The undirected GGM offers an attractive alternative to exploratory DAG estimation: the GGM is saturated rather than overidentified if all edges are present ( contains the same number of unique elements as ), does not feature equivalent models10 (there is only one unique inverse for ), does not suffer from a questionable direction of causal effect, does not require the assumption of acyclicity, and is easily parameterized using partial correlation coefficients (Epskamp et al., 2017d). These benefits come, however, at the cost of losing information on the direction of effect. To investigate the structure of a GGM under the causal model of Equation (5), in which observed variables can only be caused by other observed variables, we can invert that expression to obtain: (6) in which is still a diagonal matrix. It becomes evident that there is no longer a matrix inversion needed and that the sparsity in directly corresponds to the sparsity in ; the GGM thus acts on the same level as causal modeling. We can derive that κij equals zero if there is no directed edge between node i and j (e.g., YiYj or YiYj) and if there is no common effect of node i and node j (e.g., YiYkYj; Koller & Friedman, 2009).11 Thus, assuming a causal model as in Equation (5) generated the data, an edge in a GGM emerges as a result of a direct causal effect between the variables, or as a result of the fact that both variables have a common effect on a third variable. Note that within the causal model of Equation (5), there are no latent common causes by assumption. Edges in the GGM can therefore be indicative of potential causal effects.

A note on common effects. As mentioned above and shown in Table 1, conditioning on a common effect may induce a spurious edge in the GGM. In this case, the sign of the edge can be informative: two positive causal effects from two variables on a third lead to a negative partial correlation. As such, when observing an edge of an unexpected sign in the GGM, this may be indicative of a common effect, especially when the marginal correlation coefficient between the two variables was of the different sign. It should also be noted that conditioning on a common effect might cancel out a weak effect between two variables. In addition, because edges may be induced due to conditioning on a common effect, the GGM does not estimate a skeleton graph, a causal network with arrowheads removed (an edge may be in the GGM that is not in the causal model). Skeleton graphs can also be estimated from data (e.g., Kalisch, Maechler, & Colombo, 2017), but are not parameterized and rely on many separate conditional independence tests, potentially leading to power issues.

1.2.2. Factor models

Suppose that, instead of assuming no unobserved causes as in (5), we take the generating model of (4) and only allow for unobserved causes of the observed variables. Then, (4) reduces to the well-known factor model (Brown, 2014). The corresponding model for now becomes: which can subsequently be inverted to obtain an expression for the equivalent GGM. Golino & Epskamp (2017) provide a detailed derivation of this inverted expression and show that a factor in the factor model will lead to its indicators to cluster (all nodes connected to each other with strong edges) in the GGM. This result is in line with mathematical equivalences between factor models and network models of binary variables (Epskamp et al., in press; Kruis & Maris, 2016; Marsman et al., 2017a; Marsman, Maris, Bechger, & Glas, 2015). As there is only one unique inverse to , there is only one unique GGM for every factor model. Conversely, however, one GGM may be equivalent to many different factor models (e.g., all possible rotations of ).

Due to these equivalences, network modeling and factor modeling are closely connected. A natural first step in performing an exploratory factor analysis would be to estimate and draw a GGM model and investigate if the nodes cluster as would be expected by a factor model. Cluster-detection algorithms on the GGM could even be performed to investigate the number of factors to extract (Golino & Epskamp, 2017). Of note, however, is that many GGM estimation methods will always aim to estimate sparse GGM (i.e., contains exact zeroes), which is not expected given a factor model (except when latent variables are orthogonal). As such, estimating a sparse network does not provide evidence that a latent variable model could not have generated the data (Epskamp, Kruis, & Marsman, 2017c; Epskamp et al., in press). GGM modeling can further be used to augment factor analysis by modeling the latent variable variance-covariance matrix or the residual variance-covariance matrix as a GGM (Epskamp et al., 2017d). Modeling as a GGM leads to a latent network model, which can be used in exploratory estimation of relationships between latent variables. Modeling as a GGM leads to a residual network model, which may be used to estimate factor models while local independence is structurally violated (Epskamp et al., 2017d; Pan, Ip, & Dubé, 2017).

A note on spurious edges. When interpreting edges in the GGM as indicative of potential causal effects, it is important to note that edges in a GGM may also result from latent variables. Such edges are termed spurious, and cannot be accounted for unless the latent variable is explicitly modeled (e.g., by using the residual network model described above). The same problem occurs in exploratory DAG estimation, in which case a latent variable may induce a directed edge in the causal network. Furthermore, such spurious associations may arise in any statistical model, to the extent that unmeasured latent variables are involved. Here, the downside that GGM loses information on direction of effect turns into an upside: when an edge is indicative of a causal effect, GGMs do not retrieve the direction of effect, however, when an edge is spurious due to the influence of a latent variable, the GGM also does not introduce a strong causal hypothesis on what would happen under intervention.

2. Estimating GGMs from different sources of data

2.1. Data with independent cases

A GGM can be estimated in data sets where cases can be assumed to be independent. Three common examples of such data are cross-sectional data, in which every subject is only measured once on a set of response items, aggregated data, in which only one mean score per variable per subject is included in the data set, or n = 1 time-series data that feature large intervals between measurement occasions. In time-series data featuring shorter intervals, a GGM can be estimated as well; in this case, the network could be termed a contemporaneous network. However, as we argue in the next section on temporally ordered data, better methods exist that take temporal information into account in addition to modeling the contemporaneous effects in a GGM.

2.1.1. Estimation

In cross-sectional data analysis, only one observation per subject is available; thus, we cannot expect to estimate subject-specific means or GGM networks. It is typically assumed that the subjects all share the same distribution. That is, in which denotes the random response of subject P on all items. Similarly, in n = 1 time-series data we can make a similar assumption: in which denotes the random response of a subject on all items at time point T. In both cases, the full likelihood can be readily obtained, and the variance-covariance matrix can reliably be estimated using maximum likelihood estimation (MLE), least-square estimation, or Bayesian estimation.

Regularization. The MLE solution of —the precision matrix encoding a GGM—can be obtained by standardizing the inverse sample variance-covariance as per Equation (2). To obtain a sparse network, model search can be performed by iteratively adding and removing edges and fitting the corresponding GGM structure (Epskamp et al., 2017d). In recent literature, it has become increasingly popular to use regularization techniques, such as penalized MLE, to jointly estimate model structure and parameter values (Costantini et al., 2015; van Borkulo et al., 2014). The least absolute shrinkage and selection operator (LASSO; Tibshirani, 1996) has been shown to perform well in quickly estimating model structure and parameter estimates of a sparse GGM (Friedman, Hastie, & Tibshirani, 2008; Meinshausen & Bühlmann, 2006; Yuan & Lin, 2007). A particularly popular variant of LASSO is the graphical LASSO (GLASSO; Friedman et al., 2008), which directly penalizes elements of the inverse variance-covariance matrix (Witten, Friedman, & Simon, 2011; Yuan & Lin, 2007). The GLASSO algorithm is useful as it is typically faster than other GGM estimation algorithms (which conduct multiple separate regressions and then combine the results using Equation (3)), and requires only an estimate of the variance-covariance matrix rather than raw data (Epskamp & Fried, in press). LASSO utilizes a tuning parameter which can be chosen in a way that optimizes cross-validated prediction accuracy or that minimizes information criteria such as the extended Bayesian information criterion (EBIC; Chen & Chen, 2008). Estimating a GGM with the GLASSO algorithm in combination with EBIC model selection has been shown to work well in retrieving the true network structure (Epskamp, 2016; Foygel & Drton, 2010). For an introduction to this methodology aimed at empirical researchers, we refer the reader to Epskamp & Fried (in press).

Software. Several software packages allow for GGM estimation as described above. MLE can be performed in any programming language and in many statistical programs by inverting and subsequently standardizing the sample variance-covariance matrix. In the open-source statistical programming language R (R Core Team, 2017), automated procedures have been implemented in the corpcor package (Schafer et al., 2017) and the qgraph (Epskamp et al., 2012) package. The qgraph package also supports thresholding via significance testing or false discovery rates. The GLASSO algorithm is implemented in the glasso (Friedman, Hastie, & Tibshirani, 2014) and huge (Zhao et al., 2015) packages. EBIC-based tuning parameter selection using the glasso package has been implemented in the qgraph package. The huge package also allows for selection of the tuning parameter using cross validation or EBIC. The parcor package (Krämer, Schäfer, & Boulesteix, 2009) implements other LASSO variants to estimate the GGM. The BDgraph package (Mohammadi & Wit, 2015) implements a Bayesian method to estimate the undirected structure. Finally, fitting an estimated GGM to data can be done in the R packages ggm (Marchetti, Drton, & Sadeghi, 2015) and lvnet (Epskamp et al., 2017d).

2.2. Temporally ordered data of a single subject

In line with a call for more intraindividual and person-based research (Molenaar, 2004), an increasingly popular form of data pertains to n = 1 time series, in which a single individual is measured repeatedly over a period of time. One such situation is in clinical practice (Epskamp et al., 2018; Kroeze et al., 2017), where a patient can be measured several times per day over a period of a few weeks. We will limit our discussion to data obtained in a relatively short time-frame so that we can reasonably assume the model will remain stable over time. Then, we can apply the methodology above to obtain a GGM for the n = 1 data. However, such an analysis does not take temporal ordering of data into account (i.e., relationships between measurement occasions) and only investigates contemporaneous relationships between variables (e.g., within the same measurement occasion). This is important for several reasons. First, valuable information, especially in the context of dynamical relationships, might be contained at the temporal level rather than at the contemporaneous level. Second, not taking temporal ordering into account might bias the estimated contemporaneous relationships (see Section 4 of the supplementary materials). For example, if one variable causes itself and another variable at the next time, then not taking temporal ordering into account turns that variable into a latent cause, which would produce an edge in the GGM. Third, temporal information is needed when constructing the joint likelihood over time (e.g., to obtain the information retained in a system over time; Epskamp, 2017a; Quax, Kandhai, & Sloot, 2013). Finally, temporal information can aide in distinguishing reciprocal and cyclic effects by regarding these as acyclic effects unfolding over time.

Vector Auto-regression. The simplest way to deal with temporal ordering of cases is to incorporate the effect between consecutive measurements (Chatfield, 2016; Hamilton, 1994; Shumway & Stoffer, 2010). This is called a Lag-1 model because it includes both measurements at the current time point t as well as measurements from the previous time point t − 1. We will focus our discussion on Lag-1 models, noting that everything below also generalizes to more complicated models (e.g., Lag-2 models). In intraindividual analysis, VAR (Brandt & Williams, 2007; Rosmalen, Wenting, Roest, de Jonge, & Bos, 2012) has gained substantive footing in visualizing temporal information through networks. The lag-1 VAR model can be denoted as a regression model on the previous measurement occasion: (7) The model matrix encodes temporal predictive effects from variables on variables in the next measurement occasion, and can be used to obtain a directed network, which we term the temporal network. The variance-covariance matrix can be inverted () to obtain a GGM modeling effect within the same measurement occasion, after controlling for temporal effects. These can be displayed again as a network, which we term the contemporaneous network.

Temporal networks. Temporal networks, encoded by , have grown popular in recent psychological literature (e.g., Bringmann et al., 2013, 2015; Bos et al., 2017; Klippel et al., 2017; Snippe et al., 2017; Wigman et al., 2015). A temporal network is formed by combining a lagged variable yt − 1 and current variable yt into a single node, connected with directed edges which are weighted according to the regression parameters contained in .12 Thus, an edge in the temporal network indicates that a node predicts another node (or itself in the common case of self-loops) at the next measurement occasion, after controlling for all other variables at the previous measurement occasion. Temporal prediction is central to the concept of Granger causality in the economic literature (Eichler, 2007; Granger, 1969), and it satisfies at least the temporal requirement for causation (i.e., the cause must precede the effect). Temporal networks may thus highlight potential causal pathways. While temporal networks are typically cyclic, they can also be interpreted as summarizing a DAG unfolding over time.

Contemporaneous networks. In addition to temporal effects, VAR analyses also include contemporaneous effects, which can be modeled as a GGM. We will term this modeling framework (a VAR model with contemporaneous effects explicitly modeled and portrayed as a GGM) graphical VAR (GVAR; Wild et al., 2010).13 A useful equivalent way to denote a GVAR model is by using a conditional Gaussian distribution: Which is equivalent to Equation (7). Now, it becomes evident that if consecutive cases can be assumed to be independent, and thus , the GVAR model is exactly the same as the GGM model described above for independent cases. Thus, the GVAR model can be seen as a generalization of the GGM model to temporally ordered data. GVAR only differs from regular VAR in that the contemporaneous structure is modeled and represented as a GGM, instead of being saturated. This leads to a strikingly different interpretation of the VAR model; the VAR model can be seen as an inclusion of temporal effects on a GGM.

Temporal and contemporaneous information. Figure 2 shows a hypothetical example of the two network structures obtained in a GVAR analysis and shows how they might plausibly differ. The left panel shows the temporal network. The self-loop shows that whenever the subject in question felt energetic (or tired), this person also felt more (or less) energetic in the next measurement. The temporal network also shows us that after exercising, this person felt less energetic. The contemporaneous network in the right panel shows a plausible reverse relationship: Whenever this person exercised, he or she felt more energetic in the same measurement occasion. In psychology, there will likely be many causal relationships that occur much faster than the lag interval of a typical ESM study; in this case, these pathways will be captured in the contemporaneous network. For example, if someone is experiencing bodily discomfort, that will immediately negatively affect that person’s ability to enjoy him or herself (Epskamp et al., 2018). Especially when the measurement is on blocks of time (e.g., “since the last measurement did you feel...”), such effects are likely to be caught in the contemporaneous network.

Figure 2. A hypothetical example of two network structures obtained from a GVAR analysis. The network on the left indicates the temporal network, demonstrating that a variable predicts another variable at the next time point. The network on the right indicates the contemporaneous network, demonstrating that two variables predict each other at the same time point.

2.2.1. Estimation

Estimating saturated (fully connected temporal and contemporaneous networks) GVAR models is straightforward. First, one needs to estimate temporal effects of a regular VAR model by performing multivariate multiple regression of all variables on the previous measurement occasion, or by estimating univariate models for every variable, in which denotes the ith row of . Next one can invert the variance-covariance matrix of the residuals to obtain a GVAR model. Step-wise model selection in latent network models (Epskamp et al., 2017d) can also be used to estimate sparse GVAR models. Missing data can be handled in default ways of SEM or regression models (e.g., listwise deletion and full-information maximum likelihood), or by using more sophisticated techniques such as Bayesian estimation (Schuurman, Grasman, & Hamaker, 2016b) or the Kalman filter (Harvey, 1990; Kim & Nelson, 1999).

Novel estimation methods. A promising recent method for estimating VAR models is the Bayesian dynamical SEM implementation in version 8 of Mplus (Asparouhov, Hamaker, & Muthén, 2017; Muthén & Muthén, 2017), which includes handling of missing data, measurement invariance, and latent variables. Mplus can be used to estimate saturated GVAR models, and to perform model selection in the temporal network of a GVAR model. Model selection in the contemporaneous network of a GVAR model is not yet implemented, but credibility intervals around contemporaneous effects can be obtained by manually inverting each sampled residual variance-covariance matrix (these can be stored using the BPARAMETERS option).

When estimating GVAR models, regularization methods can be used similar to the estimation of GGMs on nontemporally ordered data. Abegaz and Wit (2013) proposed to apply LASSO estimation to jointly estimate the temporal and contemporaneous network structures using the multivariate regression with the covariance estimation (MRCE) algorithm described by Rothman, Levina, & Zhu (2010). MRCE involves iteratively optimizing , using cyclical-coordinate descent, and , using the GLASSO algorithm (Friedman et al., 2008, Friedman et al., 2014). EBIC model selection can be used to obtain the best performing model. This methodology has been implemented in two open source R packages: sparseTSCGM (Abegaz & Wit, 2015), which aims to estimate the model on repeated multivariate genetic data, and graphicalVAR (Epskamp, 2017c), which was designed to estimate the model on the psychological data of a single subject. The graphicalVAR package also allows for unregularized multivariate estimation.

An alternative to estimating GVAR models is to estimate structural VAR (SVAR; Chen et al., 2011) models, also called unified SEM (Gates, Molenaar, Hillary, Ram, & Rovine, 2010). In SVAR, the contemporaneous effects are modeled using a directed network instead of an undirected network. The sparsity of the undirected GVAR contemporaneous network corresponds in the same way to the sparsity of the directed contemporaneous network in SVAR as how the GGM corresponds to causal models (edges arise in the GGM due to edges in the causal network or conditioning on common effects). The temporal SVAR network is sparser than the temporal GVAR network, as contemporaneous mediators can be controlled for in SVAR but not in GVAR. A saturated SVAR model can be obtained by using regressions on the previous time-point as mentioned above, followed by transforming the contemporaneous variance-covariance matrix (e.g., by using a Cholesky decomposition on its inverse; Lütkepohl, 2005) and subsequently transforming the temporal effects to take contemporaneous mediators into account. This technique of obtaining an SVAR model leads to multiple solutions (Beltz & Molenaar, 2016). Step-wise model selection can also be used to estimate sparse SVAR, for example, by using model selection in (unified) SEM (Gates et al., 2010) or Bayesian dynamical SEM models (Asparouhov et al., 2017; Muthén & Muthén, 2017).

2.3. Temporally ordered data of multiple subjects

A type of data that is increasingly common due to the emergence of ESM studies is time series of multiple subjects (e.g., Bringmann et al., 2013, 2015; Mõttus et al., 2017; Schmiedek, Lövdén, & Lindenberger, 2010; Wigman et al., 2015). Such data sets pose a promising gateway to study both intraindividual dynamics and between-subjects overlap as well as their differences. Here, we assume that the number of time points might differ per person and that measurement occasions are nested in people. We can model the temporal data of every person with an individual GVAR model: in which indicates the stationary mean vector of subject p (which enters the model because we can no longer assume within-subject means are zero without loss of generality), encodes the person-specific temporal network, and encodes the person-specific contemporaneous GGM.

Multilevel modeling. To gain insight in the general network structure over subjects, we can investigate the individual networks at a second level. Doing so is termed multilevel modeling, explained in more detail in Section 2.1 of the supplementary materials. Let and encode the expected temporal and contemporaneous network when selecting a person at random. Furthermore, we can assume without loss of generality that data are grand-mean centered. We then obtain: Here, and now encode the average parameters in the population: the fixed effects. Deviations from these fixed effects, such as , are often called random effects. Besides the individual network structures, researchers often aim to estimate the structure and parameters of these fixed effects because these tell us something about the average intraindividual effect. Researchers also aim to estimate the variance-covariance structure of the random effects because it tells us something about individual differences (Bringmann et al., 2013).

The random effects can be modeled by assuming a second level normal distribution on all the parameters. This can be complicated, however, especially when modeling partial correlation coefficients in such a way (e.g., any hierarchical model for needs to take into account that this matrix must remain positive definite). The interpretation of, for example, correlations between different temporal or contemporaneous edges is also difficult. Therefore, we only focus here on a subset of the parameters where we can easily interpret the second-level model: the mean structure. As a result, if a multivariate normal is assumed for all parameters, then it is also assumed for the marginal distribution of the means—regardless of other parameters: Again, we can invert the variance-covariance matrix to obtain a GGM, which we will term the between-subjects network, a network between stationary means of different subjects.14 As such, estimating the GVAR model on n > 1 time-series analysis allows for the separation of variance into three distinct network structures: temporal networks, contemporaneous networks, and the between-subjects network.

2.3.1. Estimation

In this section, weTable 2outline several different ways in which individual network structures as well as fixed effect network structures may be estimated. We first discuss applying the methodology of estimating n = 1 GVARs discussed above to both pooled data as well as data of each subject separately, followed by a discussion of different multilevel estimation procedures that take clustering of the data into account. An overview of these methods is also included in Table 2.

Table 2. Three methods of estimating GVAR models with n > 1 subjects.

Pooled and individual LASSO estimation. First, we can estimate a GVAR model for every subject to obtain subject-specific estimates for the temporal and contemporaneous networks. Similarly, we can estimate fixed-effects networks by estimating a GVAR model on the entire within-subjects centered data set, using the sample means of every subject on every variable as a plug-in for the within-subject means. Consequently, we can estimate the between-subjects network by estimating a GGM on the sample means of each subject on all variables. We can readily apply the LASSO regularization methods described earlier for this purpose: the methodology outlined by Abegaz and Wit (2013) to estimate temporal and contemporaneous networks and the methodology outlined by Foygel & Drton (2010) to estimate the between-subjects GGM. We term this framework pooled and individual LASSO estimation and have implemented it in the R package graphicalVAR (Epskamp, 2017c). The performance of pooled and individual LASSO estimation is assessed in simulations reported in Section 3 of the supplementary materials.

Multilevel estimation. The second and third procedures described in Table 2 make use of multilevel modeling (Hamaker, 2012). Two main benefits of this approach are (1) instead of estimating the VAR model in each subject, only the fixed effects and variance-covariance of the random effects need to be estimated, and (2) afterward, estimates of subject-specific parameters can be obtained, which are somewhat pulled together (termed shrinkage). Shrinkage allows the estimation of the model for one subject to borrow information from other subjects. Multilevel estimation can be performed by specifying the multivariate model using hierarchical Bayesian Monte Carlo sampling methods or by integrating over the distribution of the random effects (Gelman & Hill, 2006; Schuurman et al., 2016b).

Multivariate Bayesian multilevel. Bayesian multivariate estimation has proven to be powerful in estimating multivariate multilevel models, especially given its flexibility in adding measurement error, latent variables, and in handling missing data (Schuurman, Houtveen, & Hamaker, 2015). Recently, the dynamic SEM methodology implemented in Mplus version 8 (Asparouhov et al., 2017; Muthén & Muthén, 2017) has made estimation of multivariate multilevel VAR models much faster and more user-friendly than other Bayesian software routines. Specifying a temporal VAR model with correlated random effects is straightforward and relatively fast to compute with a moderate number of variables (e.g., 6). At the time of writing, Mplus does not return partial correlations by default, but these can be obtained by using the BPARAMETERS option and manually inverting the sampled variance-covariance matrices. Mplus allows for specifying random effects on the contemporaneous covariances and thus, by extension, allows for estimating random contemporaneous networks in addition to random temporal networks. Specifying such a model can be done by specifying dummy latent variables for the residual covariance between each pair of variables (a prior guess on the sign of the covariance is needed). Doing so, however, can significantly increase computation length especially when all random effects are allowed to correlate. To facilitate estimation, we have implemented a function generating Mplus code for a multilevel GVAR model and subsequently running the model using the MplusAutomation package (Hallquist & Wiley, 2017) in version 0.4 of the mlVAR package, which can be called using estimator = ''Mplus'' and requires the Mplus program to be installed.

Two-step multilevel VAR. A downside of multivariate estimation is that the number of random effect covariances to be estimated increases quadratically with the number of variables. Forcing random effects to be uncorrelated helps, but places strict assumptions on the model. Bringmann et al. (2013) proposed to estimate multilevel VAR models using univariate models instead, using a frequentist estimation procedure. In this work, the multilevel VAR model is estimated by sequentially estimating univariate multilevel regression models of one variable given all lagged variables. Doing so ignores several correlations of random effects because many parameters are not estimated in the same model, simplifying the analysis: only correlations between incoming edges to the same node and the intercept of that node are included in an univariate model. This method scales up well to approximately eight variables when estimating correlated random effects and around 20 variables when estimating orthogonal random effects (or by using a moving window approach; Bringmann et al., 2015). Of note, when specifying orthogonal random effects not all random effects are assumed to be uncorrelated, merely the ones used in the same univariate model.

The methodology of Bringmann et al. (2013) does not estimate contemporaneous or between-subjects networks. To this end, we extended the algorithm in a framework we term two-step multilevel VAR. The details of this estimation procedure are explained in Section 2 of the supplementary materials. In short, we extend the methodology of Bringmann et al. (2013) by within-subject centering and by adding subject sample means as between-subjects predictors (as discussed by, e.g., Curran & Bauer, 2011; Hoffman & Stawski, 2009; Hamaker & Grasman, 2014). This allows us to estimate between-subjects networks by collecting regression coefficients as in Equation (3) and symmetrizing the resulting matrix.15 In a second step, we take the residuals of the first analysis and again perform sequential univariate multilevel regression models to predict each residual from all other residuals in the same measurement occasion. Again, these can be collected, as in Equation (3), and symmetrized to obtain contemporaneous networks. Networks can be thresholded by removing all effects that are not significant. For the between-subjects and contemporaneous networks, this results in two p-values for every edge—either both can be required to be significant (“and”-rule) or an edge can be included if one of the two p-values is significant (“or”-rule). Using the “and”-rule means erring more on the side of caution (sparser network), whereas using the “or”-rule means erring more on the side of discovery. We have implemented two-step multilevel VAR in the mlVAR package, which can be called using estimator = ''lmer'' (the default).

Choosing the estimation method. The choice of which estimator to use is not trivial and depends on the interests of the researcher. In Table 2, we list some pros and cons of each of the methodologies. In particular, multilevel estimation can be very complicated and is harder in high-dimensional settings. Assuming normally distributed parameters can also be problematic because doing so imposes that subjects cannot differ on the structure of the networks, merely on their parameterization. When a parameter (e.g., a temporal edge) is zero in some subjects but nonzero in others, then this parameter cannot be normally distributed (the distribution would peak at 0). Therefore, it is currently hard to estimate differently structured individual networks (different edges set to be exactly 0 between subjects) in multilevel estimation. Nonetheless, multilevel estimation particularly shines in that when estimating an individual network, researchers can borrow information from other subjects. We have performed simulation studies to assess the performance of the two proposed methods in this paper: pooled and individual LASSO estimation and two-step multilevel VAR. We report the results of these studies in Section 3 of the supplementary materials, which shows that both methods adequately detect the true fixed-effect network structures with increasing sample size. Having more time points per subject helps to estimate the contemporaneous and temporal networks, and having more subjects helps to estimate the between-subject networks. Two-step multilevel VAR performs well in estimating intraindividual networks when the number of observations is low, but does not perform subject-specific model selection: all estimated intra-individual networks are saturated and contain all edges. Pooled and aggregated LASSO estimation does estimate the structure of intraindividual networks, but performs poorer in intraindividual parameter estimation with fewer observations as no information is borrowed from other subjects.

GIMME. Finally, when analyzing n > 1 data, another option is to estimate SVAR models instead. A promising estimation procedure to estimate such models over many individuals, while dealing with potential heterogeneity, is “group iterative multiple model estimation” (GIMME; Gates & Molenaar, 2012), which is implemented in R using the gimme package (Lane, Gates, Molenaar, Hallquist, & Pike, 2016). In GIMME, no multilevel structure is imposed and subject-specific networks are allowed to differ in structure. Information from other subjects is borrowed, however, in that the structure of individual networks can be based on other subjects (e.g., an edge can be included because it is present in many other subjects). No shrinkage is induced on the parameter estimates that are nonzero (as would be the case in multilevel or hierarchical Bayesian modeling). A variant of GIMME that estimates the GVAR or a combination of structural and GVAR models has not yet been developed, and we note that this may be a promising avenue for further research.

3. Interpreting GGMs estimated from between-subjects data

This paper describes the estimation of network models on data from different subjects (both cross-sectional as well as person-wise average scores). Cross-sectional network modeling is often criticized for inappropriately taking cross-sectional results to be reflective of within-person causal processes (e.g., Bos & Wanders, 2016; Bos et al., 2017), as it can be shown that such results will not equal within-person processes except under strong assumptions (Molenaar, 2004). To this end, this section discusses the interpretation of GGMs estimated from data of different subjects. We first argue that cross-sectional data may be interpreted as a between-subjects analysis, assuming that between-subjects variance is dominant, and next discuss the interpretation of potential causal effects at the between-subjects level.

3.1. Cross-sectional data analysis

Within- and between-subjects variation. A type of data to which the GGM is currently often applied is data belonging to multiple subjects that are all measured only once (e.g., Isvoranu et al., 2017; van Borkulo et al., 2015). Such a data set is often termed cross-sectional data, and such an analysis is often termed a between-subjects analysis. However, the term between-subjects analysis might not be warranted, as it is difficult to distinguish between within-subject variation around an individual’s stable mean and between-subject variation of such stable within-subject means using only cross-sectional data (Hamaker, 2012). It is well known that subjects might respond differently when measured multiple times (Lord, Novick, & Birnbaum, 1968). As such, the single observation per subject leads to the time point and the subject being random: . We might make the argument that two distinct sources of variation cause the outcome (Bolger & Laurenceau, 2013). Repeated measures of a subject (here p) are distributed according to a unique within-subject model: That is, of a particular response, the subject’s score is a composite of the average stationary score and random deviation.16 These average stationary scores also differ in the population. Thus, we need to model the average stationary scores of a random subject P with a separate distribution: in which we can assume, without loss of generality, an overall mean of . We can invert the variance-covariance matrix to obtain a GGM: This GGM corresponds to a between-subjects network. The matrix can also be inverted and standardized to a GGM to obtain a within-subject network: We will term this network a within-subjects network.

The value of a cross-sectional analysis. It is immediately clear that with only one response per subject we cannot hope to estimate subject-specific variance-covariance matrices (and as a result individual GGMs). Moreover, even if we assume that within-subject effects are equal across subjects (denoted with below), this still leaves us without an estimable model because is also assumed to be normally distributed. The co-variation between responses thus becomes an unidentified blend of and : A and B may correlate in cross-sectional data because people who score on average high on A also score on average high on B (trait-level variation in ), or because when people deviate from their average on A they also tend to deviate from their average on B (state-level variation in ). Even when within- and between-subjects effects are assumed not to correlate, the GGM estimated on such data becomes which is not a simple function of the between-subjects GGM and the within-subjects GGM. Only when no short-term within-subject variation, , or no between-subjects variation, , is assumed does the cross-sectional GGM correspond exactly to one of the two networks.

Cross-sectional data analysis thus cannot disentangle between-subjects relationships from short-term within-subjects relationships (Hamaker, 2012). For example, cross-sectional analysis cannot distinguish whether or not fatigue and concentration correlate because whenever people feel fatigued they also concentrate poorly (a within-subjects effect) or because people who are on average fatigued also tend to concentrate poorly on average (a between-subjects effect). However, preliminary simulation results show that the resulting cross-sectional GGM generally does not contain edges that are not present in either the within- or between-subjects network (Epskamp, 2018). Depending on the ratio of within-to-between person variance, the cross-sectional analysis will pick up the within-subject network, the between-subject network, or a mixture of the two. As such, if one assumes between-subject variance to be dominant, the cross-sectional results may be interpreted as mainly reflecting between-subjects relations.

Cross-sectional analysis as between-subjects analysis. An important consideration is that a typical cross-sectional questionnaire or interview is vastly different than a typical ESM questionnaire, and many cross-sectional studies aim to measure variables that are more stable over time and for which a time-series analysis might not make sense. Good examples of this are recent network analyses in the area of schizophrenia (Isvoranu, Borsboom, van Os, & Guloksuz, 2016; Isvoranu et al., 2017), in which the impact of environmental factors (e.g., childhood trauma, urbanization) on psychotic symptoms and general psychopathology was studied. Such variables do not vary much over time; therefore, a cross-sectional analysis seems more suitable here. Other examples are questionnaires asking participants to rate symptoms over a period of several weeks or to describe themselves as “I am a person who...” In such cases, the cross-sectional network may be interpreted as a between-subjects network, of which we discuss the interpretation below. Note that some research indicates that state-level variance (how a person feels at the moment) does influence self-reported scores given trait-level (how a person feels on average) instructions (Brose, Lindenberger, & Schmiedek, 2013), indicating that this interpretation of cross-sectional results should be taken on a case-by-case basis depending on the topic studied.

3.2. Within- and between-subjects effects

In contrast to prior work on multilevel VAR modeling (e.g., Bringmann et al., 2013, 2015; Pe et al., 2015; Wigman et al., 2015), in this paper between-subjects effects are conceptualized in addition to the within-subjects effects in a separate GGM. Furthermore, in contrast to prior work, cross-sectional networks are not interpreted to be reflective of within-subject effects, but rather to potentially reflect a between-subjects structure, assuming that observed scores are not dominated by state-like variance (but see Brose et al., 2013). This raises the question on how such models could be interpreted. In particular, if edges in the GGM are interpreted as generating hypotheses to potential causal pathways, the question is raised how such causal effects can occur at the between-subjects level. This section therefore discusses the topic of causation at the between-subjects level. Here, we interpret the stationary means as being locally stationary: the average of a subject in a relatively short time span of measurement (e.g., a few weeks). As such, we do not interpret the mean vector as a lifetime average. Instead, we assume it could change, potentially due to experimental intervention. As a result, we argue that the between-subjects network can also be indicative of potential causal pathways—regardless of whether it is estimated from a cross-sectional interview concerning variables that are not expected to vary much over time or obtained from estimating the means from time-series data. To simplify the argumentation below, we do not discuss separate temporal and contemporaneous networks but only general within-subjects networks (a GGM of within-subject data without taking temporal ordering into account).Figure 3

Simpson’s paradox. Hamaker (2012) described an example of how within- and between-subject effects can strongly differ from each other. Suppose we let people write several texts, and we measure the number of spelling errors they make and the number of words per minute they type (typing speed). We would expect to see the seemingly paradoxical network structures shown in Figure 3, Panel (a). We would expect a positive relationship in the within-subjects network (e.g., typing faster than your average leads to making more errors). Conversely, we would expect a negative relationship in the between-subject network (e.g., people who type fast, on average, generally make fewer spelling errors). This is because people who type fast, on average, are likely to be more skilled in writing (e.g., a courtroom stenographer) and are less prone to make a lot of spelling errors, compared to someone who types infrequently. Panel (b) of Figure 3 shows another example in which the structures might differ (Hoffman 2015; provided by Hamaker 2017). These network structures show that when people exert more physical activity than their average they likely experience an elevated heart rate, while people who on average are often physically active likely have a lower average heart rate. Such a different effect depending on the level of analysis is well known in the statistical literature as Simpson’s paradox (Simpson, 1951).

Figure 3. Two hypothetical examples of differing within- and between-subject networks. The networks on the left indicates the within-subject network, showing that personal deviations from the means predict each other at the same time point, and the networks on the right indicates the between-subjects network, showing how the means of different subjects relate to one another. (a) Example based on Hamaker (2012). (b) Example based on Hoffman (2015); Hamaker (2017).

Interventionist accounts of causation. The different ways of thinking about the effects of manipulations in time-series models can be organized in terms recently developed from interventionist accounts of causation (Woodward, 2005). According to Woodward, causation is fleshed out in terms of interventions: X is a cause of Y if an intervention (natural or experimental) on X leads to a change in Y. Statistically, the interventionist account is compatible with, for example, Pearl’s 2000 semantics in terms of a “do-operator.” Here, an intervention on X is represented as Do(X = x), and the causal effect on Y is formally expressed as . Pearl distinguished this from the classical statistical association, in which no intervention is present, and we get the ordinary regression . This notation is useful here, because it can be used to show how different kinds of causal manipulations, each at the intraindividual level, can produce a signal in either the between-subjects or the within-subjects network.

Cashing out causal effects in terms of interventions is useful for understanding the intervention Do(X = x). We can think of this in terms of a random shock to the system, which sets X to value x at a particular time point and evaluates the effect on another variable Y shortly afterward. If we want to gauge this type of causal relationship, we might look at the within-subjects VAR model. Consider Hamaker’s (2012) example regarding typing errors: If a researcher forced a person to type very fast, that researcher would need to evaluate the within-subject data, which would show a positive association between typing speed and the number of errors. In this example, between-subjects data would be misleading because individual differences would probably yield a negative correlation between speed and accuracy—faster typists are more likely to make less errors.

Interventions at the mean level. However, we can also think of a manipulation that sets X to value x in a different way, for instance, by inducing a long-term change in the system that leads it to converge on X = x in expectation. To evaluate the effect of this type of intervention, it is important to consider the behavior of the system as it relates to the changes of the intercept of X. When analyzing time-series data gathered in a relatively short time-span, the within-subjects VAR network as discussed here cannot represent the relevant effects, because it assumes stationarity. However, such effects will be visible in the between-subjects network, which may thus contain important clues to the behavior of the system under potential changes in the intercept of one variable. In terms of Panel (b) of Figure 3, if we are interested in the effect of changing someone structurally—reducing the heart rate of a person on average—our preferred source of hypothesis generation would likely stem from the between-subjects model, as the corresponding within-subject model using the methods described in this paper only models deviations from the stationary mean. Such hypotheses could then be further investigated by using experimental design or lengthier longitudinal data analysis.

Many such examples can be envisioned, especially in the field of psychopathology. For instance, short-term deviations from the mean in abusing a substance might not immediately develop tolerance or lead to one suffering from work or life inferences, but a subject who abuses a substance on average over a long time period might develop these problems (example based on variables used by Rhemtulla et al., 2016). A between-subjects network could similarly show that loneliness mediates the effect of losing a spouse on depressive symptoms (Fried et al., 2015) or highlight the possible effects of childhood trauma and urbanization on psychotic symptoms (Isvoranu et al., 2016, 2017)—both cases in which within-subjects networks based on short-term deviations from the average seem less applicable. This analysis is important because it shows that, even though relevant causal interventions in psychology will typically operate at the intra-individual level, evidence for the effect of such interventions may arise at either the within- or the between-subjects level depending on the nature of the intervention.

4. Empirical examples

4.1. Reanalysis of Mõttus et al. (2017)

We reanalyzed the data of Mõttus et al. (2017) to provide an empirical example of the multilevel VAR methods described above. These data consist of two independent ESM samples, in which items tapping three of the five Five-Factor Model (neuroticism, extraversion, and conscientiousness; McCrae & John, 1992) domains were administered, as was an additional question that asked participants how much they had exercised since the preceding measurement occasion. Sample 1 consisted of 26 people providing 1323 observations in total, and Sample 2 consisted of 62 people providing a total of 2193 observations. Participants in Sample 1 answered questions three times per day, whereas participants in Sample 2 answered questions five times per day. In both samples, the minimum time between measurements was 2 hours. For more information about the samples and the specific questions asked, we refer readersFigure 4to Mõttus et al. (2017).

To obtain an easier and more interpretable example, we first only analyzed questions aimed as measuring the extraversion trait and the question measuring exercise. This leads to five variables of interest: questions pertaining to feeling outgoing, energetic, adventurous, or happy and the question measuring participants’ exercise habits. We analyzed the data using the two-step multilevel VAR procedure as described in detail in the Section 2 of the supplementary materials. We used the mlVAR package, version 0.4, for the estimation of this model. Because the number of variables was small, we estimated the model using correlated temporal and contemporaneous random effects. We ran the model separately for both samples and computed the fixed effects for the temporal, contemporaneous, and between-subjects networks. Correlations of the edge weights indicated that all three networks showed high correspondence between the two samples (temporal network: 0.82, contemporaneous network: 0.94, between-subjects network: 0.70). Owing to the degree of replicability, we combined the two samples and estimated the model on the combined data.

Results. Figure 4 shows the estimated fixed effects of the temporal, contemporaneous, and between-subjects network. In these figures, only significant edges (α = 0.05) are shown. In the contemporaneous and between-subjects networks, an edge was retained if one of the two regressions on which the partial correlation is based was significant (the so-called “or” rule; van Borkulo et al., 2014). These results are in line with the hypothetical example shown in Figure 2: People who exercised were more energetic while exercising and less energetic after exercising. In the between-subjects network, no relationship between exercising,Figure 5, Figure 6feeling energetic, and feeling adventurous was found. The between-subjects network, however, showed a strong relationship between feeling adventurous and exercising: People who, on average, exercised more also felt, on average, more adventurous. This relationship was not present in the temporal network and much weaker in the contemporaneous network. Also noteworthy is that people were less outgoing after exercising. Figure 5 shows the standard deviation of the random effects in the temporal and contemporaneous networks. The largest individual differences in the temporal network were found in the auto-regressions, and the largest individual differences in the contemporaneous network were found in relationship between exercising and feeling energetic.

Figure 4. The estimated fixed effects of the three network structures obtainable in multilevel VAR. The model is based on ESM data of 88 people providing a total of 3516 observations. Due to differences in the scale of the networks, the temporal network was drawn with a different maximum value (i.e., the value indicating the strongest edge in the network) than the contemporaneous and between-subjects networks. Edges that were not significantly different from zero were removed from the networks.

Figure 5. The networks showing the standard deviation of random effects in the temporal and contemporaneous networks. Due to scale differences, networks were plotted using different maximum values.

In addition to using only the extraversion and exercise items, we also ran the model on all 17 administered items in the data set. In this analysis, we used orthogonal random effects to estimate the model because correlated random effects cannot be estimated with such a large number of variables. Figure 6 shows the estimated fixed effects of the three network structures; it can be seen that indicators of the three traits tend to cluster together in all three networks. Regarding the node exercise, we found the same relationships between exercise, energetic, and adventurous (also found in the previous example) in the larger networks. Furthermore, we noted that exercising was connected to feeling angry in the between-subjects network but not in the other networks. Finally, there was a between-subjects connection between exercising and feeling self-disciplined: People who, on average, exercised more also felt, on average, more self-disciplined.

Figure 6. The estimated fixed effects of the three network structures based on all 17 variables administered. Only significant edges are shown. Legend: 1 = “Worried”; 2 = “Organized”; 3 = “Ambitious”; 4 = “Depressed”; 5 = “Outgoing”; 6 = “Self-Conscious”; 7 = “Self-Disciplined”; 8 = “Energetic”; 9 = “Frustrated”; 10 = “Focused”; 11 = “Guilty”; 12 = “Adventurous”; 13 = “Happy”; 14 = “Control”; 15 = “Achieved”; 16 = “Angry”; 17 = “Exercise.”

4.2. Reanalysis of Bringmann et al. (2013)

To showcase additional information that can be obtained using the GGM model, we reanalyzed the data set used and made publicly available by Bringmann et al. (2013), which has been collected by Geschwind et al. (2011). This data set contains ESM measures of 129 participants, which was collected in two periods over 6 days each: a baseline period and a posttreatment period (mindfulness treatment and a control group). Participants answered 60 measurements per period. Similar to Figure 1 of Bringmann et al. (2013), we analyzed only the baseline data set on the six items selected by Bringmann et al. (2013). We estimated the networks using three modeling frameworks discussed in Table 2. First, we analyze data using multilevel Bayesian estimation using Mplus version 8 (model generated using the mlVAR package). We estimated correlated random effects for the temporal effects but only fixed effects for the contemporaneous effects (making these random led to slow convergence). The model was estimated using three chains that ran until convergence. Nights were handled by adding a row of missing values between consecutive days. Second, we analyzed the data using two-step multilevel VAR estimation as implemented in the mlVAR package, using an “and”-rule and estimating correlated random temporal and contemporaneous effects. Finally, we estimated the data using pooled and individual LASSO estimation using the graphicalVAR package, using γ = 0.25. In the final two analyses, we did not regress the first measurement of the day on the last measurement of the previous day, and removed all pairs of lagged and current variables that contained missing responses. The final sample size was 5927 observations. Edges were retained if they were significant at the α = 0.05 level, or if 0 was not included in the 95% credibility interval.

Results. Figure 7 shows the resulting network structures, and shows that all three methods are mostly aligned. Unsurprisingly, the temporal networks are very similar to those reported by Bringmann et al. (2013).17 Both the temporal and contemporaneous network are in line with what would be expected under a unidimensional auto-correlated latent variable model (many edges selected, low-rank structure, edges of expected sign) with the exception of the positive temporal edge from “fearful” to “pleasant” in the two-step multilevel network (which was not selected by the other methods). Of note is that Bayesian multilevel estimation resulted in a sparser temporal network. This difference in sparsity is possibly because the multivariate Bayesian multilevel approach more accurately represents the uncertainties in parameter estimation, while two-step multilevel VAR estimates the model piecewise and pooled LASSO estimation does not take the multilevel structure into account. Remarkable is the positive edge between “sad” and “relaxed” in the two-step multilevel between-subjects network, which is based on two significant positive Level 2 regression coefficients (β = 0.202, p = 0.046 and β = 0.151, p = 0.036) where the estimated between-subjects correlation is strongly negative (− 0.53). This edge is especially remarkable since both nodes are strongly connected to other nodes in the network. The Bayesian multilevel between-subjects network showed a similar positive edge between “cheerful” and “worry.” This is noteworthy because under a unidimensional factor model, we would not expect partial correlation coefficients to switch sign from marginal correlation coefficients (Holland & Rosenbaum, 1986; van Bork, Grasman, & Waldorp, 2016). A possible way the partial correlation coefficient switches sign is if it has been conditioned on one or more common effects between the two variables of interest (in this case, potentially “worry,” “pleasant,” or “fearful”). Of course, these effects must be interpreted with great care, especially given the high p-values; we did not control for multiple comparisons, and the same edges are not retained in the other methods. Still, it is noteworthy that if this edge is weak or nonexistent, the between-subjects structure is still not in line with a unidimensional factor model. In such a factor model, “sad” and “relaxed”(which feature the most connections) would be expected to have a strong negative edge between them (a depression factor would lead to “sad” having a strong positive factor loading and “relaxed” having a strong negative factor loading).

Figure 7. Reanalysis of the Geschwind et al. (2011) data set used by Bringmann et al. (2013). (a) Fixed effect network structures estimated via multilevel Bayesian estimation. (b) Fixed effect network structures estimated via two-step multilevel estimation. (c) Fixed effect network structures estimated via pooled and individual LASSO estimation.

5. Discussion

We discussed the Gaussian graphical model (GGM; Lauritzen, 1996), an undirected network model of partial correlation coefficients, and discussed its utility in the analysis of psychological data sets. The GGM presents a promising exploratory data analysis tool that allows for different levels of interpretation: (1) Edges in the GGM can be interpreted without reliance on a causal interpretation and merely used to show which variables predict each other. (2) Causal effects between variables result in an edge, whereas the lack of a causal effect results in no edge, except in the presence of latent variables or a common effect. The GGM can, therefore, be seen as hypothesis generating structures that highlight potential causal pathways. (3) Undirected models can be used and interpreted as causal data-generating process and have been used as such in several fields of research.

The GGM can readily be estimated on any data set that contains multiple observations of the same variables (e.g., multiple people in cross-sectional data or multiple responses in time-series data). LASSO regularization methods perform especially well in estimating such a GGM structure. In temporally ordered data (e.g., n = 1 time series), the graphical VAR (GVAR; Wild et al., 2010) model generalizes the GGM to incorporate temporal effects. We showed how two network structures can be obtained: a temporal network, which is a directed network of regression coefficients between lagged and current variables, and a contemporaneous network, which is a GGM describing the relationships that remain after controlling for temporal effects. In temporally ordered data of multiple subjects (e.g., n > 1 time series), the natural combination of cross-sectional and time-series data came by adding a third network structure: the between-subjects network, which is a GGM that describes relationships between the stationary means of subjects. We proposed two methods to estimate the three network structures: (1) two-step multilevel estimation, which we implemented in the open source R package mlVAR, and (2) pooled and individual VAR model estimations using LASSO regularization, which we implemented in the open source R package graphicalVAR.

5.1. Limitations and challenges

Multilevel estimation. The presented methods are not without problems and have several limitations. With regard to multilevel estimation, first, multivariate estimation of the multilevel VAR model is not yet feasible for larger data sets. As such, the proposed two-step multilevel VAR combines univariate models. Doing so, however, means that not all parameters are in the same model. In addition, univariate models do not readily provide estimates of the contemporaneous networks, which must be estimated in a second step. Second, even when multivariate estimation is possible, it is still challenging to estimate a multilevel model on contemporaneous networks due to the requirement of positive definite matrices. Third, when more than approximately eight variables are measured, estimating the multilevel models with correlated random effects is no longer feasible in open source software. In this case, orthogonal random effects can be used, which induce a level of parsimony that may not be substantively plausible. Finally, even when orthogonal estimation is used, multilevel analysis runs very slowly in models with more than 20 variables. As such, multilevel VAR analysis of high-dimensional data sets is not yet feasible. To this end, we discussed pooling within-subject centered data and estimating fixed-effects models using LASSO regularization (Abegaz & Wit, 2013). This performed on par with multilevel estimation in higher sample sizes and allows researchers to scale up the analysis. However, individual network estimation using separate VAR models does not borrow information from other subjects and performs poorly in low sample sizes. Promising developments are new LASSO methods in which shrinkage from subject-specific parameters to their mean is attained through penalization rather than hierarchical modeling (Hastie et al., 2015). Future research should investigate the utility of such models in estimating individual network structures that might differ in structure but borrow information from other subjects in its estimation.

VAR modeling assumptions. These limitations on the estimation methods come with more limitations in the statistical models themselves. VAR modeling, especially, is not without problems and faces severe challenges (Hamaker, Ceulemans, Grasman, & Tuerlinckx, 2015; Hamaker & Wichers, 2017). We made several assumptions that can be problematic. For instance, in characterizing the likelihood of time-series data, we need to assume that the conditional distribution of variables at time t given time t − 1 are the same for all t. That raises two distinct assumptions: (1) The difference in time between measurements are roughly equal, and (2) the parameters do not change over time. Equidistance in time is especially important for the interpretation of temporal networks. Promising work is being done in this area where VAR networks can be estimated on nonequidistant data sets (Driver, Oud, & Voelkle, 2017; Oravecz, Tuerlinckx, & Vandekerckhove, 2009; Oud & Jansen, 2000). The assumption of stationarity is needed to estimate structures when data are limited but might not be tenable especially in longer time series (Rovine & Walls, 2006). Promising time-varying estimation procedures are being developed (Bringmann et al., 2016; Haslbeck & Waldorp, 2016a), but are not yet extended to the GVAR framework. Furthermore, the interpretation of temporal coefficients when represented as a network is not without discussion, and several different methods for standardization exist (Bulteel, Tuerlinckx, Brose, & Ceulemans, 2016; Schuurman, Ferrer, de Boer-Sonnenschein, & Hamaker, 2016a).18

Normality. Another particularly important assumption made in this paper is that of multivariate normality. Indeed, Equation (1) makes this assumption and all other equations follow from this. The assumption of normality is not without problems (Terluin, de Boer, & de Vet, 2016). However, it is not always straightforward to deal with these issues, because violations of normality may arise for many different reasons. When data are not normally distributed, then they cannot be represented properly using only the means vector and variance-covariance matrix. As a result, the GGM does not properly characterize the joint likelihood function. When data are measured on a different scale (Stevens, 1946), a different graphical model can be used, such as the Ising model for binary data (Epskamp et al., in press; van Borkulo et al., 2014) or a mixed graphical model for categorical and Poisson-distributed variables as well as binary and Gaussian variables (Haslbeck & Waldorp, 2016b). Such models have yet to be extended to time-series analysis, especially in separating temporal and contemporaneous effects as the GVAR model does. When data are continuous but not normal, multiple reasons can (again) contribute to this. When the underlying process is normal but the measured variables are on a transformed scale, transforming data back to normal should offer a solution (Liu, Lafferty, & Wasserman, 2009), but when the process itself is nonnormal, such as skewed residuals, the entire modeling framework does not correctly capture the likelihood. Finally, multivariate normality assumes all relationships between variables are linear. When this is not the case, the GGM and VAR model (which fit linear effects) will not properly describe the data. We encourage future researchers to focus on the problem of normality and to develop new methods of overcoming these challenges.

Interpretation. Finally, it should be noted that when taking a causal interpretation of edges, all methods discussed in this paper are exploratory in nature and can only generate hypotheses—they do not confirm causal relations. The analyses showcased in this paper can also be used without relying on a causal interpretation and allow researchers to obtain insights into the predictive relationships present in the data—regardless of theory with respect to the data-generating model. Under the assumptions of multivariate normality, stationarity, and the Lag-1 factorization, the networks show how variables predict each other over time (temporal network), within time (contemporaneous network), and on average (between-subjects network). Furthermore, during the thresholding of edges in the multilevel analyses, we did not apply a correction for multiple testing by default. We deliberately chose this because our aim was to present exploratory hypothesis-generating structures, and not correcting for multiple testing yields greater sensitivity.

6. Conclusion

This paper provides a methodological overview of how the GGM can be used in various different kinds of psychological data. The GGM can be used to map out unique variance in cross-sectional data or at the contemporaneous and between-subjects levels of time-series analysis. We contrasted this method to exploratory estimation of causal models. While losing information on the direction of effect, estimating GGMs offers an attractive alternative in that these models are uniquely identified, well parameterized, closely related to causal models and also offer exploratory insight on predictive effects between observed variables. When the aim is to discover psychological dynamics, the GGM can be used as a hypothesis generating technique inspiring future research or therapy directions (Epskamp et al., 2018; Kroeze et al., 2017). For example, an effect found in a cross-sectional analysis could inspire a time-series study, a contemporaneous effect could inspire a shorter time-lag time-series study and a between-subjects effect could inspire lengthy longitudinal studies. All network structures may inspire experimental design, or to gather a mixture of observational and experimental data (Magliacane et al., 2017). The GGM thus provides a powerful addition to the exploratory toolbox in behavioral research.

Article Information

Conflict of Interest Disclosures: Each author signed a form for disclosure of potential conflicts of interest. No authors reported any financial or other conflicts of interest in relation to the work described.

Ethical Principles: The authors affirm having followed professional ethical guidelines in preparing this work. These guidelines include obtaining informed consent from human participants, maintaining ethical treatment and respect for the rights of human or animal participants, and ensuring the privacy of participants and their data, such as ensuring that individual participants cannot be identified in reported results or from publicly available original or archival data.

Funding: This work was supported by Grant 406-11-066 from the NWO.

Role of the Funders/Sponsors: None of the funders or sponsors of this research had any role in the design and conduct of the study; collection, management, analysis, and interpretation of data; preparation, review, or approval of the manuscript; or decision to submit the manuscript for publication.

Acknowledgments: The authors would like to thank Laura Bringmann, Noémi Schuurman, Oisín Ryan, and Ellen Hamaker for helpful tips and invigorating discussions, and Katharina Jorgensen for valuable comments on earlier versions of this paper. An earlier version of this paper has been adapted as a chapter in the dissertation of the main author (Epskamp, 2017b).

Supplemental material

HMBR_A_1454823_Supplementary.zip

Download Zip (3009 KB)

Additional information

Funding

This work was supported by Grant 406-11-066 from the NWO.

Notes

1 CRAN link: http://cran.r-project.org/package=mlVAR Github link (developmental): http://www.github.com/SachaEpskamp/mlVAR.

2 CRAN link: http://cran.r-project.org/package=graphicalVAR Github link (developmental): http://www.github.com/SachaEpskamp/graphicalVAR.

3 We use capitalized subscripts to denote random variables and lower case subscripts to denote fixed variables. A variable can potentially be fixed with respect to one subscript but random with respect to another. Supplementary materials’ Section 1 contains a complete overview of the notation used in this paper.

4 Because we assume data to be centered, we do not need to model the (grand) mean vector. This simplifies notation.

5 This relationship can be traced back much further. For example, Heiser (2017) traced this relationship back to the work of Guttman (1938).

6 Many publications make use of the default color setup used in qgraph (Epskamp, Cramer, Waldorp, Schmittmann, & Borsboom, 2012): green for positive edges and red for negative edges. A later version of qgraph includes the option theme = ”colorblind” using a more colorblind friendly coloring scheme and setting the positive edge color to blue. This option has been used for all graphs in this paper. Note that some publications (e.g., Schuurman, 2016) also use blue and red edges but use red to denote positive and blue to denote negative effects akin to a heat map.

7 This expression may differ by a scalar, depending on the estimation method. For example, by default R computes the variance-covariance matrix by using n − 1 in the denominator, but computes Var(ϵCi) by using nm in the denominator. This denominator is cancelled out in Equation (2) when standardizing to partial correlation coefficients.

8 This expression should not be confused with Equation (3), in which is obtained by performing univariate multiple regressions in which error terms are not independent.

9 This matrix is often denoted using the Greek letter instead of . We use here to avoid confusion with the contemporaneous variance-covariance matrix used below, which is not diagonal.

10 Note that the uniqueness of the GGM relates to the psychometric model: for every there is only one unique inverse and vice versa. When estimating a GGM from data, different estimation methods may lead to different estimated GGMs

11 A common effect node is also termed a “collider” in the literature on causal modeling.

12 Note, in graph theory it is common to encode a network using a weights matrix in which the row indicates the node of origin and the column indicates the row of destination. As such, to obtain the directed weights matrix to draw a temporal network needs to be transposed.

13 Wild et al. (2010) do not use the term graphical VAR in the exact same way we do, and use it more to refer to graphical modeling in a VAR framework, including structural VAR. We use the term here as described because having an explicit term helps in contrasting GVAR from, e.g., structural VAR.

14 Of note, it is also possible to invert and standardize the full random effects variance-covariance matrix, which would lead to different between-subjects relationships between the means as well (partial correlations after conditioning on other means and all other between-subject parameters such as edges). We do not do that here as (a) such a network is hard to interpret, and (b) most estimation methods we mention do not return the full random effects variance-covariance matrix (especially the correlations between temporal and contemporaneous edges are hard to obtain).

15 Standardizing regression parameters from nodewise multilevel models to partial correlation coefficients does not lead to perfectly identical estimates.

16 Section 4 of the supplementary materials shows that when consecutive cases (t and t + 1) are assumed dependent, such a zero-order network may result from a mixture of temporal and contemporaneous effects as described above. The discussion here does not concern estimation of model parameters and hence does not require an assumption of independence of cases.

17 The networks differ because the estimation of temporal effects differs in that measures are within-subjects centered and subject means are included as Level 2 predictors.

18 We standardized every data set before analyzing and used the standardization of Wild et al. (2010) for temporal networks in n = 1 and pooled temporal networks. GGMs are readily standardized by using partial correlation coefficients (Equation (2)), which have been used in all GGMs shown in this paper.

    References

  • Abegaz, F., & Wit, E. (2013). Sparse time series chain graphical models for reconstructing genetic networks. Biostatistics, 14(3), 586599. doi: 10.1093/biostatistics/kxt005. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Abegaz, F., & Wit, E. (2015). SparseTSCGM: Sparse time series chain graphical models (R package version 2.2). Retrieved from http://CRAN.R-project.org/package=SparseTSCGM. [Google Scholar]
  • Asparouhov, T., Hamaker, E., & Muthén, B. (2017). Dynamic structural equation models. Structural Equation Modeling: A Multidisciplinary Journal, 25(3), 359388. doi: 10.1080/10705511.2017.1406803 [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Beltz, A. M., & Molenaar, P. C. (2016). Dealing with multiple solutions in structural vector autoregressive models. Multivariate Behavioral Research, 51(2–3), 357373. doi: 10.1080/00273171.2016.1151333. [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Bolger, N., & Laurenceau, J. (2013). Intensive longitudinal methods. New York, NY, USA: Guilford. [Google Scholar]
  • Borsboom, D., & Cramer, A. O. J. (2013). Network analysis: An integrative approach to the structure of psychopathology. Annual Review of Clinical Psychology, 9, 91121. doi: 10.1146/annurev-clinpsy-050212-185608. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Borsboom, D., Cramer, A. O. J., Schmittmann, V. D., Epskamp, S., & Waldorp, L. J. (2011). The small world of psychopathology. PloS One, 6(11), e27407. doi: 10.1371/journal.pone.0027407. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Bos, E. H., & Wanders, R. B. (2016). Group-level symptom networks in depression. JAMA Psychiatry, 73(4), 411411. doi: 10.1001/jamapsychiatry.2015.3103. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Bos, F. M., Snippe, E., de Vos, S., Hartmann, J. A., Simons, C. J. P., van der Krieke, L., .. Wichers, M. (2017). Can we jump from cross-sectional to dynamic interpretations of networks? Implications for the network perspective in psychiatry. Psychotherapy and Psychosomatics, 86, 185177. doi: 10.1159/000453583. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Brandt, P. T., & Williams, J. T. (2007). Multiple time series models (vol. 148). Thousand Oaks, CA, USA: Sage Publications, Inc. [Crossref][Google Scholar]
  • Bringmann, L. F., Hamaker, E. L., Vigo, D. E., Aubert, A., Borsboom, D., & Tuerlinckx, F. (2016). Changing dynamics: Time-varying autoregressive models using generalized additive modeling. Psychological Methods, 22(3), 409425. doi: 10.1037/met0000085. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Bringmann, L. F., Lemmens, L. H., Huibers, M. J., Borsboom, D., & Tuerlinckx, F. (2015). Revealing the dynamic network structure of the Beck Depression Inventory-II. Psychological Medicine, 45(04), 747757. doi: 10.1017/S0033291714001809. [Crossref], [PubMed][Google Scholar]
  • Bringmann, L. F., Vissers, N., Wichers, M., Geschwind, N., Kuppens, P., Peeters, F., .. Tuerlinckx, F. (2013). A network approach to psychopathology: New insights into clinical longitudinal data. PLOS ONE, 8(4), e60188. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Brose, A., Lindenberger, U., & Schmiedek, F. (2013). Affective states contribute to trait reports of affective well-being. Emotion, 13(5), 940. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Brown, T. A. (2014). Confirmatory factor analysis for applied research. London, UK: Guilford Publications. [Google Scholar]
  • Bulteel, K., Tuerlinckx, F., Brose, A., & Ceulemans, E. (2016). Using raw var regression coefficients to build networks can be misleading. Multivariate Behavioral Research, 51(2–3), 330344. doi: 10.1080/00273171.2016.1150151. [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Chatfield, C. (2016). The analysis of time series: An introduction. Boca Raton, FL, USA: CRC press. [Google Scholar]
  • Chen, G., Glen, D. R., Saad, Z. S., Hamilton, J. P., Thomason, M. E., Gotlib, I. H., .. Cox, R. W. (2011). Vector autoregression, structural equation modeling, and their synthesis in neuroimaging data analysis. Computers in Biology and Medicine, 41(12), 11421155. doi: 10.1016/j.compbiomed.2011.09.004. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Chen, J., & Chen, Z. (2008). Extended Bayesian information criteria for model selection with large model spaces. Biometrika, 95(3), 759771. doi: 10.1093/biomet/asn034. [Crossref], [Web of Science ®][Google Scholar]
  • Costantini, G., Epskamp, S., Borsboom, D., Perugini, M., Mõttus, R., Waldorp, L. J., .. Cramer, A. O. J. (2015). State of the art personality research: A tutorial on network analysis of personality data in R. Journal of Research in Personality, 54, 1329. doi: 10.1016/j.jrp.2014.07.003. [Crossref], [Web of Science ®][Google Scholar]
  • Cramer, A. O., van Borkulo, C. D., Giltay, E. J., van der Maas, H. L., Kendler, K. S., Scheffer, M., .. Borsboom, D. (2016). Major depression as a complex dynamic system. PloS One, 11(12), e0167490. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Cramer, A. O. J., Sluis, S., Noordhof, A., Wichers, M., Geschwind, N., Aggen, S. H., .. Borsboom, D. (2012). Dimensions of normal personality as networks in search of equilibrium: You can’t like parties if you don’t like people. European Journal of Personality, 26(4), 414431. doi: 10.1002/per.1866. [Crossref], [Web of Science ®][Google Scholar]
  • Cramer, A. O. J., Waldorp, L., van der Maas, H., & Borsboom, D. (2010). Comorbidity: A Network Perspective. Behavioral and Brain Sciences, 33(2–3), 137150. doi: 10.1017/S0140525X09991567. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Curran, P. J., & Bauer, D. J. (2011). The disaggregation of within-person and between-person effects in longitudinal models of change. Annual Review of Psychology, 62, 583619. doi: 10.1146/annurev.psych.093008.100356. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Dalege, J., Borsboom, D., van Harreveld, F., van den Berg, H., Conner, M., & van der Maas, H. L. (2016). Toward a formalized account of attitudes: The causal attitude network (can) model. Psychological Review, 123(1), 222. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Driver, C. C., Oud, J. H. L., & Voelkle, M. C. (2017). Continuous time structural equation modelling with r package ctsem. Journal of Statistical Software, 77, 135. doi: 10.18637/jss.v077.i05. [Crossref], [Web of Science ®][Google Scholar]
  • Drton, M., & Maathuis, M. H. (2017). Structure learning in graphical modeling. Annual Review of Statistics and Its Application, 4, 365393. doi: 10.1146/annurev-statistics-060116-053803. [Crossref], [Web of Science ®][Google Scholar]
  • Eichler, M. (2007). Granger causality and path diagrams for multivariate time series. Journal of Econometrics, 137(2), 334353. doi: 10.1016/j.jeconom.2005.06.032. [Crossref], [Web of Science ®][Google Scholar]
  • Epskamp, S. (2016). Regularized Gaussian psychological networks: Brief report on the performance of extended BIC model selection. Retrieved from http://arxiv.org/abs/1606.05771. [Google Scholar]
  • Epskamp, S. (2017a). Discussion: The road ahead. In Network psychometrics, 237248. Retrieved from http://sachaepskamp.com/dissertation/Discussion.pdf. [Google Scholar]
  • Epskamp, S. (2017b). Discovering psychological dynamics. In Network psychometrics, 85114. Retrieved from http://sachaepskamp.com/dissertation/Chapter6.pdf. [Google Scholar]
  • Epskamp, S. (2017c). graphicalVAR: Graphical VAR for experience sampling data (R package version 0.1.6). Retrieved from http://CRAN.R-project.org/package=graphicalVAR. [Google Scholar]
  • Epskamp, S. (2018). Preliminary simulations on the interpretation of cross-sectional Gaussian graphical models. PsyArXiv preprint, 54xrs. doi: 10.17605/OSF.IO/54XRS. [Crossref][Google Scholar]
  • Epskamp, S., Borsboom, D., & Fried, E. I. (2017a). Estimating psychological networks and their accuracy: A tutorial paper. Behavior Research Methods, 50(1), 195212. doi: 10.3758/s13428-017-0862-1. [Crossref], [Web of Science ®][Google Scholar]
  • Epskamp, S., Cramer, A., Waldorp, L., Schmittmann, V. D., & Borsboom, D. (2012). qgraph: Network visualizations of relationships in psychometric data. Journal of Statistical Software, 48(1), 118. doi: 10.18637/jss.v048.i04. [Crossref][Google Scholar]
  • Epskamp, S., Deserno, M. K., & Bringmann, L. F. (2017b). mlVAR: Multi-level vector autoregression (R package version 0.4). Retrieved from https://CRAN.R-project.org/package=mlVAR. [Google Scholar]
  • Epskamp, S., & Fried, E. I. (in press). A tutorial on estimating regularized psychological networks. Psychological Methods. DOI: 10.1037/met0000167 [Crossref][Google Scholar]
  • Epskamp, S., Kruis, J., & Marsman, M. (2017c). Estimating psychopathological networks: Be careful what you wish for. PloS One, 12(6), e0179891. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Epskamp, S., Maris, G., Waldorp, L., & Borsboom, D. (in press). Network psychometrics. In P. Irwing, D. Hughes, & T. Booth (Eds.), Handbook of psychometrics. New York, NY, USA: Wiley. Retrieved from arxiv.org/abs/1609.02818. [Google Scholar]
  • Epskamp, S., Rhemtulla, M., & Borsboom, D. (2017d). Generalized network psychometrics: Combining network and latent variable models. Psychometrika, 82(4), 904927. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Epskamp, S., van Borkulo, C. D., Isvoranu, A. M., Servaas, M. N., van der Veen, D. C., Riese, H.,... Cramer, A. O. J. (2018). Personalized network modeling in psychopathology: The importance of contemporaneous and temporal connections. Clinical Psychological Science. doi: 10.1177/2167702617744325. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Foygel, R., & Drton, M. (2010). Extended Bayesian information criteria for Gaussian graphical models. Advances in Neural Information Processing Systems, 23, 20202028. Retrieved from https://arxiv.org/abs/1011.6640. [Google Scholar]
  • Fried, E. I., Bockting, C., Arjadi, R., Borsboom, D., Amshoff, M., Cramer, O. J., .. Stroebe, M. (2015). From loss to loneliness: The relationship between bereavement and depressive symptoms. Journal of Abnormal Psychology, 124(2), 256265. doi: 10.1037/abn0000028. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Fried, E. I., Epskamp, S., Nesse, R. M., Tuerlinckx, F., & Borsboom, D. (2016). What are ‘good’ depression symptoms? Comparing the centrality of DSM and non-DSM symptoms of depression in a network analysis. Journal of Affective Disorders, 189, 314320. doi: 10.1016/j.jad.2015.09.005. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Friedman, J. H., Hastie, T., & Tibshirani, R. (2008). Sparse inverse covariance estimation with the graphical lasso. Biostatistics, 9(3), 432441. doi: 10.1093/biostatistics/kxm045. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Friedman, J. H., Hastie, T., & Tibshirani, R. (2014). glasso: Graphical lasso-estimation of Gaussian graphical models (R package version 1.8). Retrieved from https://CRAN.R-project.org/package=glasso. [Google Scholar]
  • Gates, K. M., & Molenaar, P. C. (2012). Group search algorithm recovers effective connectivity maps for individuals in homogeneous and heterogeneous samples. NeuroImage, 63(1), 310319. doi: 10.1016/j.neuroimage.2012.06.026. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Gates, K. M., Molenaar, P. C., Hillary, F. G., Ram, N., & Rovine, M. J. (2010). Automatic search for fMRI connectivity mapping: An alternative to Granger causality testing using formal equivalences among SEM path modeling, var, and unified SEM. Neuroimage, 50(3), 11181125. doi: 10.1016/j.neuroimage.2009.12.117. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Gelman, A., & Hill, J. (2006). Data analysis using regression and multilevel/hierarchical models. New York, NY, USA: Cambridge University Press. [Crossref][Google Scholar]
  • Geschwind, N., Peeters, F., Drukker, M., van Os, J., & Wichers, M. (2011). Mindfulness training increases momentary positive emotions and reward experience in adults vulnerable to depression: a randomized controlled trial. Journal of Consulting and Clinical Psychology, 79(5), 618628. doi: 10.1037/a0024595. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Golino, H. F., & Epskamp, S. (2017). Exploratory graph analysis: A new approach for estimating the number of dimensions in psychological research. PlosOne, 2(6), e0174035. doi: 10.1371/journal.pone.0174035. [Crossref][Google Scholar]
  • Granger, C. W. J. (1969). Investigating causal relations by econometric models and cross-spectral methods. Econometrica: Journal of the Econometric Society, 37(3), 424438. [Crossref], [Web of Science ®][Google Scholar]
  • Guttman, L. (1938). A note on the derivation of formulae for multiple and partial correlation. The Annals of Mathematical Statistics, 9(4), 305308. Retrieved from http://www.jstor.org/stable/2957470. [Crossref][Google Scholar]
  • Hallquist, M., & Wiley, J. (2017). MplusAutomation: Automating Mplus model estimation and interpretation [Computer software manual] (R package version 0.7). Retrieved from https://CRAN.R-project.org/package=MplusAutomation. [Google Scholar]
  • Hamaker, E., Ceulemans, E., Grasman, R., & Tuerlinckx, F. (2015). Modeling affect dynamics: State of the art and future challenges. Emotion Review, 7(4), 316322. doi: 10.1177/1754073915590619. [Crossref], [Web of Science ®][Google Scholar]
  • Hamaker, E. L. (2012). Why researchers should think “within-person”: A paradigmatic rationale. In M. R. Mehl, & T. S. Conner (Eds.), Handbook of research methods for studying daily life (pp. 4361). New York, NY: Guilford Press. Retrieved from https://www.researchgate.net/publication/266896375. [Google Scholar]
  • Hamaker, E. L. (2017). A brief history of dynamic modeling in psychology. 82th Annual Meeting of the Psychometric Society (IMPS), Zurich: Switzerland. [Google Scholar]
  • Hamaker, E. L., & Grasman, R. P. (2014). To center or not to center? Investigating inertia with a multilevel autoregressive model. Frontiers in Psychology, 5, 1492. doi: 10.3389/fpsyg.2014.01492. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Hamaker, E. L., & Wichers, M. (2017). No time like the present. Current Directions in Psychological Science, 26(1), 1015. doi: 10.1177/0963721416666518. [Crossref], [Web of Science ®][Google Scholar]
  • Hamilton, J. D. (1994). Time series analysis (vol. 2). Princeton, NJ, USA: Princeton university press. [Google Scholar]
  • Harvey, A. C. (1990). Forecasting, structural time series models and the Kalman filter. Cambridge, UK: Cambridge university press. [Crossref][Google Scholar]
  • Haslbeck, J. M. B., & Waldorp, L. J. (2016a). mgm: Structure estimation for time-varying mixed graphical models in high-dimensional data. arXiv preprint, page arXiv:1510.06871. [Google Scholar]
  • Haslbeck, J. M. B., & Waldorp, L. J. (2016b). Structure estimation for mixed graphical models in high dimensional data. arXiv preprint, page arXiv:1510.05677. [Google Scholar]
  • Hastie, T., Tibshirani, R., & Wainwright, M. (2015). Statistical learning with sparsity: The Lasso and generalizations. Boca Raton, FL, USA: CRC Press. [Crossref][Google Scholar]
  • Hayduk, L. A. (2009). Finite feedback cycling in structural equation models. Structural Equation Modeling, 16(4), 658675. doi: 10.1080/10705510903206030. [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Heiser, W. J. (2017). Early psychometric contributions to Gaussian graphical modeling: A tribute to Louis Guttman (1916-1987). Zurich, Switzerland: 82th Annual Meeting of the Psychometric Society (IMPS). [Google Scholar]
  • Hoffman, L. (2015). Longitudinal analysis: Modeling within-person fluctuation and change. New York, NY, USA: Routledge. [Crossref][Google Scholar]
  • Hoffman, L., & Stawski, R. S. (2009). Persons as contexts: Evaluating between-person and within-person effects in longitudinal analysis. Research in Human Development, 6(2–3), 97120. doi: 10.1080/15427600902911189. [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Holland, P. W., & Rosenbaum, P. R. (1986). Conditional association and unidimensionality in monotone latent variable models. The Annals of Statistics, 14(4), 15231543. doi: 10.1214/aos/1176350174. [Crossref], [Web of Science ®][Google Scholar]
  • Ising, E. (1925). Beitrag zur theorie des ferromagnetismus. Zeitschrift für Physik A Hadrons and Nuclei, 31(1), 253258. doi: 10.1007/BF02980577. [Crossref][Google Scholar]
  • Isvoranu, A. M., Borsboom, D., van Os, J., & Guloksuz, S. (2016). A network approach to environmental impact in psychotic disorders: Brief theoretical framework. Schizophrenia Bulletin, 42(4), 870873. doi: 10.1093/schbul/sbw049. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Isvoranu, A. M., van Borkulo, C. D., Boyette, L., Wigman, J. T. W., Vinkers, C. H., Borsboom, D., & GROUP Investigators, (2017). A network approach to psychosis: Pathways between childhood trauma and psychotic symptoms. Schizophrenia Bulletin, 43(1), 187196. doi: 10.1093/schbul/sbw055. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Kalisch, M., & Bühlmann, P. (2007). Estimating high-dimensional directed acyclic graphs with the pc-algorithm. Journal of Machine Learning Research, 8(Mar), 613636. Retrieved from http://www.jmlr.org/papers/v8/kalisch07a.html. [Google Scholar]
  • Kalisch, M., Mächler, M., Colombo, D., Maathuis, M. H., & Bühlmann, P. (2012). Causal inference using graphical models with the R package pcalg. Journal of Statistical Software, 47(11), 126. doi: 10.18637/jss.v047.i11. [Crossref], [Web of Science ®][Google Scholar]
  • Kalisch, M., Maechler, M., & Colombo, D. (2017). pcalg: Estimation of CPDAG/PAG and causal inference using the IDA algorithm. (R package version 2.5-0). Retrieved from https://CRAN.R-project.org/package=pcalg. [Google Scholar]
  • Kaplan, D. (2000). Structural equation modeling: Foundations and extensions. CA, USA: Sage, Thousand Oaks. [Google Scholar]
  • Kim, C.-J., & Nelson, C. R. (1999). State-space models with regime switching: Classical and Gibbs-sampling approaches with applications. Cambridge, MA, USA: The MIT Press. [Google Scholar]
  • Klippel, A., Viechtbauer, W., Reininghaus, U., Wigman, J., van Borkulo, C., MERGE,.. Wichers, M. (2017). The cascade of stress: A network approach to explore differential dynamics in populations varying in risk for psychosis. Schizophrenia Bulletin, 44(2), 328337. doi: 10.1093/schbul/sbx037. [Crossref], [Web of Science ®][Google Scholar]
  • Koller, D., & Friedman, N. (2009). Probabilistic graphical models: Principles and techniques. Cambridge, MA, USA: MIT Press. [Google Scholar]
  • Kossakowski, J. J., Epskamp, S., Kieffer, J. M., van Borkulo, C. D., Rhemtulla, M., & Borsboom, D. (2015). The application of a network approach to health-related quality of life (HRQoL): Introducing a new method for assessing hrqol in healthy adults and cancer patient. Quality of Life Research, 25, 781792. doi: 10.1007/s11136-015-1127-z. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Krämer, N., Schäfer, J., & Boulesteix, A.-L. (2009). Regularized estimation of large-scale gene association networks using graphical Gaussian models. BMC Bioinformatics, 10(1), 124. doi: 10.1186/1471-2105-10-384. [Crossref], [PubMed][Google Scholar]
  • Kroeze, R., Van Veen, D., Servaas, M., Bastiaansen, J., Oude Voshaar, R., Borsboom, D., .. Riese, H. (2017). Personalized feedback on symptom dynamics of psychopathology: A proof-of-principle study. Journal for Person-Oriented Research, 3(1), 110. doi: 10.17505/jpor.2017.01. [Crossref][Google Scholar]
  • Kruis, J., & Maris, G. (2016). Three representations of the ising model. Scientific Reports, 6, 34175. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Lane, S., Gates, K., Molenaar, P., Hallquist, M., & Pike, H. (2016). Gimme: Group iterative multiple model estimation. (R package version 0.1-7). Retrieved from https://CRAN.R-project.org/package=gimme. [Google Scholar]
  • Lauritzen, S. L. (1996). Graphical models. Oxford, UK: Clarendon Press. [Google Scholar]
  • Liu, H., Lafferty, J. D., & Wasserman, L. (2009). The nonparanormal: Semiparametric estimation of high dimensional undirected graphs. The Journal of Machine Learning Research, 10, 22952328. Retrieved from https://arxiv.org/abs/0903.0649. [Web of Science ®][Google Scholar]
  • Lord, F. M., Novick, M. R., & Birnbaum, A. (1968). Statistical theories of mental test scores. Oxford, UK: Addison-Wesley. [Google Scholar]
  • Lütkepohl, H. (2005). New introduction to multiple time series analysis. Berlin, Germany: Springer-Verlag. [Crossref][Google Scholar]
  • MacCallum, R. C., Wegener, D. T., Uchino, B. N., & Fabrigar, L. R. (1993). The problem of equivalent models in applications of covariance structure analysis. Psychological Bulletin, 114(1), 185199. doi: 10.1037/0033-2909.114.1.185. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Magliacane, S., van Ommen, T., Claassen, T., Bongers, S., Versteeg, P., & Mooij, J. M. (2017). Causal transfer learning. arXiv preprint arXiv:1707.06422. [Google Scholar]
  • Marchetti, G. M., Drton, M., & Sadeghi, K. (2015). ggm: Functions for graphical Markov models (R package version 2.3). Retrieved from https://CRAN.R-project.org/package=ggm. [Google Scholar]
  • Marsman, M., Borsboom, D., Kruis, J., Epskamp, S., van Bork, R., Waldorp, L.,... Maris, G. (2017a). An introduction to network psychometrics: Relating ising network models to item response theory models. Multivariate Behavioral Research, 53(1), 1535. doi: 10.1080/00273171.2017.1379379. [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Marsman, M., Maris, G., Bechger, T., & Glas, C. (2015). Bayesian inference for low-rank ising networks. Scientific Reports, 5(9050), 17. doi: 10.1038/srep09050. [Crossref], [Web of Science ®][Google Scholar]
  • Marsman, M., Waldorp, L., & Maris, G. (2017b). A note on large-scale logistic prediction: Using an approximate graphical model to deal with collinearity and missing data. Behaviormetrika, 44(2), 513534. doi: 10.1007/s41237-017-0024-x. [Crossref][Google Scholar]
  • McCrae, R. R., & John, O. P. (1992). An introduction to the five-factor model and its applications. Journal of Personality, 60(2), 175215. doi: 10.1111/j.1467-6494.1992.tb00970.x. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • McNally, R. J., Robinaugh, D. J., Wu, G. W., Wang, L., Deserno, M. K., & Borsboom, D. (2015). Mental disorders as causal systems a network approach to posttraumatic stress disorder. Clinical Psychological Science, 3(6), 836849. doi: 10.1177/2167702614553230. [Crossref], [Web of Science ®][Google Scholar]
  • Meehl, P. E. (1990). Why summaries of research on psychological theories are often uninterpretable. Psychological Reports, 66(1), 195244. doi: 10.2466/pr0.1990.66.1.195. [Crossref], [Web of Science ®][Google Scholar]
  • Meinshausen, N., & Bühlmann, P. (2006). High-dimensional graphs and variable selection with the lasso. The Annals of Statistics, 34(3), 14361462. doi: 10.1214/009053606000000281. [Crossref], [Web of Science ®][Google Scholar]
  • Mohammadi, A., & Wit, E. C. (2015). BDgraph: An R package for Bayesian structure learning in graphical models. arXiv preprint, page arXiv:1501.05108. Retrieved from https://CRAN.R-project.org/package=BDgraph. [Google Scholar]
  • Molenaar, P. C. (2004). A manifesto on psychology as idiographic science: Bringing the person back into scientific psychology, this time forever. Measurement, 2(4), 201218. doi: 10.1207/s15366359mea0204_1. [Taylor & Francis Online][Google Scholar]
  • Mõttus, R., Epskamp, S., & Francis, A. (2017). Within-and between individual variability of personality characteristics and physical exercise. Journal of Research in Personality, 69, 139148. doi: 10.1016/j.jrp.2016.06.017. [Crossref], [Web of Science ®][Google Scholar]
  • Murphy, K. P. (2012). Machine learning: A probabilistic perspective. Cambridge, MA, USA: MIT Press. [Google Scholar]
  • Muthén, L. K., & Muthén, B. O. (2017). Mplus user’s guide: Statistical analysis with latent variables. Statistical analysis with latent variables. Version. Retrieved from https://www.statmodel.com/download/usersguide/MplusUserGuideVer_8.pdf. [Google Scholar]
  • Myin-Germeys, I., Oorschot, M., Collip, D., Lataster, J., Delespaul, P., & van Os, J. (2009). Experience sampling research in psychopathology: Opening the black box of daily life. Psychological Medicine, 39(09), 15331547. doi: 10.1017/S0033291708004947. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Oravecz, Z., Tuerlinckx, F., & Vandekerckhove, J. (2009). A hierarchical Ornstein–Uhlenbeck model for continuous repeated measurement data. Psychometrika, 74(3), 395418. doi: 10.1007/s11336-008-9106-8. [Crossref], [Web of Science ®][Google Scholar]
  • Oud, J. H., & Jansen, R. A. (2000). Continuous time state space modeling of panel data by means of SEM. Psychometrika, 65(2), 199215. doi: 10.1007/BF02294374. [Crossref], [Web of Science ®][Google Scholar]
  • Pan, J., Ip, E. H., & Dubé, L. (2017). An alternative to post hoc model modification in confirmatory factor analysis: The Bayesian lasso. Psychological Methods, 22(4), 687. doi: 10.1037/met0000112. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Pe, M. L., Kircanski, K., Thompson, R. J., Bringmann, L. F., Tuerlinckx, F., Mestdagh, M., .. Jonides, J. (2015). Emotion-network density in major depressive disorder. Clinical Psychological Science, 3(2), 292300. doi: 10.1177/2167702614540645. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Pearl, J. (2000). Causality: Models, reasoning, and inference. New York, NY: Cambridge University Press. [Google Scholar]
  • Pourahmadi, M. (2011). Covariance estimation: The glm and regularization perspectives. Statistical Science, 26(3), 369387. doi: 10.1214/11-STS358. [Crossref], [Web of Science ®][Google Scholar]
  • Quax, R., Kandhai, D., & Sloot, P. M. (2013). Information dissipation as an early-warning signal for the Lehman brothers collapse in financial time series. Scientific Reports, 3, 1898. doi: 10.1038/srep01898. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • R Core Team, (2017). R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing. Retrieved from https://www.R-project.org/. [Google Scholar]
  • Rhemtulla, M., Fried, E. I., Aggen, S. H., Tuerlinckx, F., Kendler, K. S., & Borsboom, D. (2016). Network analysis of substance abuse and dependence symptoms. Drug and Alcohol Dependence, 161, 230237. doi: 10.1016/j.drugalcdep.2016.02.005. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Rigdon, E. E. (1995). A necessary and sufficient identification rule for structural models estimated in practice. Multivariate Behavioral Research, 30(3), 359383. doi: 10.1207/s15327906mbr3003_4. [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Rosmalen, J. G., Wenting, A. M., Roest, A. M., de Jonge, P., & Bos, E. H. (2012). Revealing causal heterogeneity using time series analysis of ambulatory assessments: application to the association between depression and physical activity after myocardial infarction. Psychosomatic Medicine, 74(4), 377386. doi: 10.1097/PSY.0b013e3182545d47. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Rothman, A. J., Levina, E., & Zhu, J. (2010). Sparse multivariate regression with covariance estimation. Journal of Computational and Graphical Statistics, 19(4), 947962. doi: 10.1198/jcgs.2010.09188. [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Rovine, M. J., & Walls, T. A. (2006). Multilevel autoregressive modeling of interindividual differences in the stability of a process. In Theodore A. Walls & Joseph L. Schafer (Eds.), Models for Intensive Longitudinal Data (pp. 124147). Oxford, UK: Oxford University Press. doi: 10.3389/fpsyg.2017.00262. [Crossref][Google Scholar]
  • Schafer, J., Opgen-Rhein, R., Zuber, V., Ahdesmaki, M., Silva, A. P. D., & Strimmer, K. (2017). corpcor: Efficient estimation of covariance and (partial) correlation (R package version 1.6.9). Retrieved from https://CRAN.R-project.org/package=corpcor. [Google Scholar]
  • Schmiedek, F., Lövdén, M., & Lindenberger, U. (2010). Hundred days of cognitive training enhance broad cognitive abilities in adulthood: Findings from the cogito study. Frontiers in Aging Neuroscience, 2, 27. doi: 10.3389/fnagi.2010.00027. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Schmittmann, V. D., Cramer, A. O. J., Waldorp, L. J., Epskamp, S., Kievit, R. A., & Borsboom, D. (2013). Deconstructing the construct: A network perspective on psychological phenomena. New Ideas in Psychology, 31(1), 4353. doi: 10.1016/j.newideapsych.2011.02.007. [Crossref], [Web of Science ®][Google Scholar]
  • Schuurman, N. K. (2016). Measurement error and person-specific reliabilities in multilevel autoregressive modeling. In Multilevel autoregressive modeling in psychology: Snags and solutions, 139178. Retrieved from http://www.nkschuurman.com/NKSchuurman_dissertation.pdf. [Google Scholar]
  • Schuurman, N. K., Ferrer, E., de Boer-Sonnenschein, M., & Hamaker, E. L. (2016a). How to compare cross-lagged associations in a multilevel autoregressive model. Psychological Methods, 21(2), 206. doi: 10.1037/met0000062. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Schuurman, N. K., Grasman, R. P. P. P., & Hamaker, E. L. (2016b). A comparison of inverse-Wishart prior specifications for covariance matrices in multilevel autoregressive models. Multivariate Behavioral Research, 51(2–3), 185206. doi: 10.1080/00273171.2015.1065398. [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Schuurman, N. K., Houtveen, J. H., & Hamaker, E. L. (2015). Incorporating measurement error in n= 1 psychological autoregressive modeling. Frontiers in Psychology, 6. doi: 10.3389/fpsyg.2015.01038. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Scutari, M. (2010). Learning Bayesian networks with the bnlearn R package. Journal of Statistical Software, 35(3), 122. doi: 10.18637/jss.v035.i03. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Shumway, R. H., & Stoffer, D. S. (2010). Time series analysis and its applications: With R examples. New York, NY, USA: Springer Science & Business Media. doi: 10.1007/978-3-319-52452-8. [Crossref][Google Scholar]
  • Simpson, E. H. (1951). The interpretation of interaction in contingency tables. Journal of the Royal Statistical Society. Series B (Methodological), 13(2), 238241. Retrieved from http://www.jstor.org/stable/2984065. [Crossref], [Web of Science ®][Google Scholar]
  • Snippe, E., Viechtbauer, W., Geschwind, N., Klippel, A., De Jonge, P., & Wichers, M. (2017). The impact of treatments for depression on the dynamic network structure of mental states: Two randomized controlled trials. Scientific Reports, 7, 46523. doi: 10.1038/srep46523. [Crossref], [PubMed][Google Scholar]
  • Stevens, S. S. (1946). On the theory of scales of measurement. Science, New Series, 103(2684), 677680. doi: 10.1126/science.103.2684.677. [Crossref], [Web of Science ®][Google Scholar]
  • Terluin, B., de Boer, M. R., & de Vet, H. C. (2016). Differences in connection strength between mental symptoms might be explained by differences in variance: Reanalysis of network data did not confirm staging. PloS One, 11(11), e0155205. doi: 10.1371/journal.pone.0155205. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society. Series B (Methodological), 58, 267288. Retrieved from http://www.jstor.org/stable/2346178. [Crossref], [Web of Science ®][Google Scholar]
  • van Bork, R., Grasmans, R. P. P. P., & Waldorp, L. J. (2016). Unidimensional factor models imply weaker partial correlations than zero-order correlations. arXiv preprint, in press. page arXiv:1610.03375. [Google Scholar]
  • van Borkulo, C. D., Borsboom, D., Epskamp, S., Blanken, T. F., Boschloo, L., Schoevers, R. A., .. Waldorp, L. J. (2014). A new method for constructing networks from binary data. Scientific Reports, 4(5918), 110. doi: 10.1038/srep05918. [Crossref][Google Scholar]
  • van Borkulo, C. D., Boschloo, L., Borsboom, D., Penninx, B. W. J. H., Waldorp, L. J., & Schoevers, R. A. (2015). Association of symptom network structure with the course of depression. JAMA Psychiatry, 72(12), 12191226. doi: 10.1001/jamapsychiatry.2015.2079. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • van der Maas, H. L., Dolan, C. V., Grasman, R. P., Wicherts, J. M., Huizenga, H. M., & Raijmakers, M. E. (2006). A dynamical model of general intelligence: The positive manifold of intelligence by mutualism. Psychological Review, 113(4), 842861. doi: 10.1037/0033-295X.113.4.842. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Wigman, J., van Os, J., Borsboom, D., Wardenaar, K., Epskamp, S., Klippel, A., .. Wichers, M. (2015). Exploring the underlying structure of mental disorders: Cross-diagnostic differences and similarities from a network perspective using both a top-down and a bottom-up approach. Psychological Medicine, 45(11), 23752387. doi: 10.1017/S0033291715000331. [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Wild, B., Eichler, M., Friederich, H.-C., Hartmann, M., Zipfel, S., & Herzog, W. (2010). A graphical vector autoregressive modeling approach to the analysis of electronic diary data. BMC Medical Research Methodology, 10(1), 28. doi: 10.1186/1471-2288-10-28. [Crossref], [PubMed][Google Scholar]
  • Witten, D. M., Friedman, J. H., & Simon, N. (2011). New insights and faster computations for the graphical lasso. Journal of Computational and Graphical Statistics, 20(4), 892900. doi: 10.1198/jcgs.2011.11051a. [Taylor & Francis Online], [Web of Science ®][Google Scholar]
  • Woodward, J. (2005). Making things happen: A theory of causal explanation. Oxford, UK: Oxford University Press. Retrieved from http://www.jstor.org/stable/3840611. [Google Scholar]
  • Wright, S. (1921). Correlation and causation. Journal of Agricultural Research, 20(7), 557585. [Google Scholar]
  • Yuan, M., & Lin, Y. (2007). Model selection and estimation in the Gaussian graphical model. Biometrika, 94(1), 1935. doi: 10.1093/biomet/asm018. [Crossref], [Web of Science ®][Google Scholar]
  • Zhao, T., Li, X., Liu, H., Roeder, K., Lafferty, J., & Wasserman, L. (2015). huge: High-dimensional undirected graph estimation (R package version 1.2.7). Retrieved from https://CRAN.R-project.org/package=huge [Google Scholar]

Appendix: Glossary of terms