Longitudinal Analysis of the ABCD® Study

Authors

Samuel W. Hawes

Andrew K. Littlefield

Daniel A. Lopez

Kenneth J. Sher

Erin L. Thompson

Raul Gonzalez

Laika Aguinaldo

Ashley R. Adams

Mohammadreza Bayat

Amy L. Byrd

Luis FS Castro-de-Araujo

Anthony Dick

Steven F. Heeringa

Christine M. Kaiver

Sarah M. Lehman

Lin Li

Janosch Linkersdörfer

Thomas J. Maullin-Sapey

Michael C. Neale

Thomas E. Nichols

Samantha Perlstein

Susan F. Tapert

Colin E. Vize

Margot Wagner

Rebecca Waller

Wesley K. Thompson

Abstract

The Adolescent Brain Cognitive Development® (ABCD) Study provides a unique opportunity to investigate developmental processes in a large, diverse cohort of youths, aged approximately 9-10 at baseline and assessed annually for 10 years. Given the size and complexity of the ABCD Study, researchers analyzing its data will encounter a myriad of methodological and analytical considerations. This review provides an examination of key concepts and techniques related to longitudinal analyses of the ABCD Study data, including: 1) characterization of the factors associated with variation in developmental trajectories; 2) assessment of how level and timing of exposures may impact subsequent development; 3) quantification of how variation in developmental domains may be associated with outcomes, including mediation models and reciprocal relationships. We emphasize the importance of selecting appropriate statistical models to address these research questions. By presenting the advantages and potential challenges of longitudinal analyses in the ABCD Study, this review seeks to equip researchers with foundational knowledge and tools to make informed decisions as they navigate and effectively analyze and interpret the multi-dimensional longitudinal data currently available.

1 Introduction

The Adolescent Brain Cognitive Development (ABCD) Study® is the largest longitudinal investigation of neurodevelopment and child health in the United States. Conceived and initiated by the National Institutes of Health (NIH), this landmark prospective longitudinal study aims to transform our understanding of the genetic and environmental factors impacting neurodevelopment and their roles in behavioral and health outcomes across ten years of adolescence Volkow et al. (2018). At its heart, the study is designed to chart the course of human development across multiple interacting domains from late childhood to early adulthood and to identify factors that lead to both positive and negative outcomes. Central to achieving these goals is the commitment of the ABCD Study and its NIH funders to an open science framework, intended to facilitate sharing of data and analytical methods by espousing practices that increase access, integrity, and reproducibility of scientific research. In this context, the ABCD Study is a collaboration with the broader research community.

The size and scope of the ABCD Study data allow the research community to perform a large variety of developmental analyses of both substantive and methodological interest, presenting a unique opportunity to significantly advance our understanding of how a multitude of biopsychosocial processes unfold across critical periods of development. In this paper, we describe models and methods for longitudinal analysis of ABCD Study data that can address these fundamental scientific aims, including: 1) characterization of the genetic and environmental factors associated with variation in developmental trajectories; 2) assessment of how the level and timing of exposures may impact subsequent neurodevelopment; 3) quantification of how variation in developmental domains may be associated with outcomes, including mediation models and reciprocal relationships. We instantiate these longitudinal analyses in worked examples using the ABCD Release 5.1 data with accompanying R scripts. Worked examples are available in Quarto files, accessible in the project’s GitHub repository.

1.1 The ABCD Study Data

The ABCD Study enrolled a cohort of n=11,880 participants born between 2006-2008 and aged approximately 9-10 years at baseline, each with a parent/guardian. The study sample was recruited from households in defined catchment areas for each of the 21 (originally 22) study sites across the United States. Information regarding funding agencies, recruitment sites, investigators, and project organizations can be obtained at https://abcdstudy.org. The ABCD Study design is described in more detail in Garavan et al. (2018) and Dick et al. (2021).

The ABCD Study is currently collecting longitudinal data on a rich variety of outcomes that will enable the construction of complex statistical models, potentially incorporating factors from many domains. Each new wave of data collection provides another building block for characterizing developmental trajectories and implementing longitudinal analyses that allow researchers to characterize normative development, to identify variables that presage deviations from normative development, and to assess a range of variables associated with biopsychosocial outcomes of interest. These data include: 1) a neurocognitive battery Luciana et al. (2018); Wesley K. Thompson et al. (2019); 2) mental and physical health assessments Barch et al. (2018); 3) measures of culture and environment Gonzalez et al. (2021); Zucker et al. (2018); 4) substance use Lisdahl et al. (2021); 5) gender identity and sexual health Potter et al. (2022); 6) biospecimens Uban et al. (2018); 7) structural and functional brain imaging Casey et al. (2018); Hagler et al. (2019); Palmer et al. (2022); 8) geolocation-based environmental exposure data Fan et al. (2021); 9) wearables and mobile technology Bagot et al. (2018); and 10) whole-genome genotyping Loughnan et al. (2020). Many of these measures are collected at in-person annual visits, with brain imaging collected at baseline and every other year going forward. A limited number of assessments are collected in semi-annual brief telephone or online assessments.

Data are publicly released approximately annually, currently through the NIMH Data Archive (NDA). The study’s earliest data releases consisted primarily of one or two visits per participant. However, the most recent public release as of the writing of this paper (Release 5.1) contains data collected across five annual visits, including three brain imaging assessments (baseline, year 2 follow-up, and year 4 follow-up visits) for at least a subset of the cohort. Hence, starting with Release 5.0, it is feasible for researchers to begin focusing on the characterization of neurodevelopmental and other trajectories.

1.2

• Part I. Introduction

  • The ABCD Study®

• Part II. Developmental Research

  • Fundamental Concepts

• Part III. Longitudinal Analysis

  • Methods & Analysis

• Part IV. Supplemental materials

  • Linked open-source resources

2 Developmental Research

2.1 Basic Concepts and Considerations

There are several important concepts to consider when conducting longitudinal analyses in a developmental context. These include different ways of thinking about the developmental course, whether certain periods of development are relatively sensitive or insensitive to various types of insults or stressors, whether some time periods or situations inhibit the expression of individual differences due to extreme environmental pressures, and whether the same behavior manifested at different times represents the same or different phenomena.

Moreover, in the case of developmentally-focused longitudinal research, each new measurement occasion not only provides a more extended portrait of the child’s life course but also brings with it greater methodological opportunities to make use of statistical models that distinguish within- from between-person effects and that loosen constraints that need to be imposed on the furtherance of critical scientific questions.

For example, collecting two or more within-person observations on the same construct at different times enables estimation of individual rates of change (slopes) where more observations allow for more precise estimates of individual slopes (random slopes), as well as characterization of non-linear development. Rate of change or other trajectory characteristics may be more informative about individuals than the simple snapshots of level differences that cross-sectional data are limited to informing about. Cross-sectional age-related differences across individuals are poor substitutes for longitudinal trajectory estimates, except under highly restrictive assumptions, e.g., parallel trajectories and lack of age, cohort and experience effects Wesley K. Thompson et al. (2011). Appreciation of these and other issues can help to guide the analysis and interpretation of data and aid translation to clinical and public health applications.

2.1.1 Vulnerable periods.

Adolescent development progresses normatively from less mature to more mature levels of functioning. However, unique epochs and experiences can alter the course of this idealized form of development. Consider research that shows cannabis use during adolescence is associated with later psychosis to a greater degree than cannabis use initiated later in development Arseneault et al. (2002); Bechtold et al. (2016); Hasan et al. (2020); Semple, McIntosh, and Lawrie (2005). Similarly, rodent brains are especially sensitive to the neurotoxic effects of alcohol on brain structure and learning early in development, corresponding to early adolescence in humans Spear (2016); Crews et al. (2000); Ji et al. (2018). In another example, longitudinal data from the National Consortium on Alcohol and Neurodevelopment in Adolescence (NCANDA) show that binge drinking is associated more strongly with decrements in gray matter volume early in adolescence compared to later Infante et al. (2022). These examples highlight the importance of considering the role of vulnerable periods – e.g., temporal windows of rapid brain development or remodeling during which the effects of environmental stimuli on the developing brain may be particularly pronounced– when trying to establish an accurate understanding of the association between exposures and outcomes.

2.1.2 Developmental disturbances.

Whereas vulnerable periods heighten neurobiological susceptibility to environmental influences, at other times, environmental exposures will tend to suppress stability and disrupt the orderly stochastic process of normative development (e.g., Schulenberg et al. (2019)). This situation reflects a developmental disturbance in that the normal course of development is “altered” for a time by some time-limited process. In such cases, we might find that prediction of behavior in the period of the disturbance is reduced and/or, similarly, the behavior exhibited during the disturbance might have less predictive power with respect to distal outcomes compared to the behavior exhibited before and following the disrupted period. That is, once the environmental pressures are removed (or the individual is removed from the environment), patterns of individual differences (and autoregressive effects) recover to levels similar to those prior to entering the environment.

2.1.3 Developmental snares and cascade effects.

Normative development can also be upended by experiences (e.g., drug use) that, through various mechanisms, disrupt the normal flow of development wherein each stage establishes a platform for the next. For instance, substance use could lead to association with deviant peers, precluding opportunities for learning various adaptive skills and prosocial behaviors, in effect creating a “snare” that delays psychosocial development, such as maturing out of adolescent antisocial behavior Moffitt (2015). Relatedly, the consequences of these types of events can cascade (e.g., school dropout, involvement in the criminal justice system) so that the effects of the snare are amplified (e.g., Masten et al. (2005); Rogosch, Oshri, and Cicchetti (2010)). Although conceptually distinct from vulnerable periods, both types of developmental considerations highlight the importance of viewing behavior in the context of development and attempting to determine how various developmental pathways unfold. Longitudinal data are crucial in this context to assess individual levels of development prior to and following onset of experiences or other environmental factors (e.g., the ABCD Study collected data starting at ages 9-10 and hence before the onset of substance use for the vast majority of participants).

2.1.4 Mediational Processes.

Questions regarding the biological mechanisms whereby exposures impact outcomes can often be framed in terms of mediation analyses MacKinnon, Fairchild, and Fritz (2007); VanderWeele (2016). Mediation analyses can be implemented using the causal steps approach Baron and Kenny (1986) and structural equation models (SEM) Preacher, Zhang, and Zyphur (2011). More recently, mediation models have been adapted for longitudinal exposures, mediators, and/or outcomes Bind et al. (2016); VanderWeele and Tchetgen Tchetgen (2017). All of these modeling approaches decompose the total effects of an exposure on an outcome into direct and indirect effects, where indirect effects of an exposure flow through its impact on a mediating process. VanderWeele and Tchetgen Tchetgen (2017) details conditions under which the direct and indirect causal effects can be in a longitudinal setting. An important example of mediational analyses in the ABCD Study is the impact of exposures on behavioral outcomes (e.g., neurocognition, mental health, substance use) via their impact on the brain, as quantified by imaging-derived phenotypes (IDPs). Methods for mediational analyses using multi-dimensional IDPs have been developed and applied to functional MRI data Lindquist (2012); Zhao et al. (2018).

3 Longitudinal Data

3.1 Considerations and Challenges

The hallmark characteristic of longitudinal data analysis (LDA) is the administration of repeated measurements of the same constructs on assessment targets (e.g., individuals, families) across time. The primary rationale for collecting longitudinal data is to assess within-person change over time, allowing researchers to estimate individual developmental trajectories and the genetic and person-level factors that may impact these trajectories. Administering repeated measurements more frequently or over longer periods enables researchers to ask more nuanced questions and to make stronger inferences.

3.1.1 Two Time Points versus Three or More.

Although the clear leap from cross-sectional to the realm of longitudinal data involves going from one assessment to two or more assessments, there are also notable distinctions in designs based on two-assessment points versus three or more measurement occasions. Just as cross-sectional data can be informative in some situations, two waves of data can be beneficial in contexts such as when an exposure is involved (e.g., pre/post tests), or if the central goal is prediction (e.g., trying to predict scores on Variable A at time T as a function of prior scores on Variable A and Variable B at time T-1). At the same time, analyses of data based on two assessments are inherently limited on multiple fronts. As Rogosa, Brandt, and Zimowski (1982) noted over forty years ago, “Two waves of data are better than one, but maybe not much better” (p. 744).

These sentiments are reflected in more contemporary recommendations regarding best-practice guidelines for prospective data, which increasingly emphasize the benefits of additional measurement occasions for trajectory estimation, model identification and accurate parameter inferences. This is also consistent with recommendations that developmental studies include three or more assessment points, given it is impossible for data based on two-time points to determine the shape of development (given that linear change is the only estimable form for two assessment waves; (see Duncan and Duncan (2009)). Research designs that include three (but preferably more) time points allow for non-linear trajectory estimation and increasingly nuanced analyses that more adequately tease apart sources of variation and covariation among the repeated assessments King et al. (2018)– a key aspect of developmental research.

To illustrate, developmental theories are useful for understanding patterns of within-individual change over time (discussed in further detail, below); however, two data points provide meager information on change at the person level. This point is further underscored in a recent review of statistical models commonly touted as distinguishing within-individual vs between-individual sources of variance in which the study authors concluded “… researchers are limited when attempting to differentiate these sources of variation in psychological phenomenon when using two waves of data” and perhaps more concerning, “…the models discussed here do not offer a feasible way to overcome these inherent limitations” Andrew K. Littlefield et al. (2021). It is important to note, however, that despite the current focus on two-wave designs versus three or more assessment waves, garnering three assessment points is not a panacea for longitudinal modeling. Indeed, several contemporary longitudinal models designed to isolate within-individual variability (e.g., the Latent Curve Model with Structured Residuals [LCM: SR]; Patrick J. Curran et al. (2014)) require at least four assessments to parameterize fully and, more generally, increasingly accurate and nuanced parameter estimates are obtained as more assessment occasions are used Duncan and Duncan (2009).

3.1.2 Types of stability and change

If one were to try to sum up what developmental trajectories in a living organism are exactly, one could plausibly argue they are the patterns of stability and change in its phenotypes as the organism traverses the life course. Symbolically, developmental trajectories can be expressed as fi(t), a possibly multivariate function of time t, specific to the ith individual and typically taking values in the real numbers for continuous phenotypes and the integers for discrete phenotypes. Ideally, t is a biologically meaningful temporal index (e.g., calendar age) as opposed to an exogenous progression of events (e.g., study visit number). Properties of interest might include rate of change over time, degree of smoothness (e.g., continuously differentiable), shape (e.g., polynomial or asymptotic behavior), how and how much f(t) differs across individuals, and what factors predict either within-individual variation (at different times) or between-individual variation (either overall or at specific times).

There are a few different ways to think about patterns of stability and change (see Figure 1). Consider measuring school disengagement at the start of middle school and the end of middle school. A common first step may be to compare sixth graders’ average disengagement values and eighth graders’ disengagement values. This comparison of the average scores for the same group of individuals at multiple time points is referred to as “mean-level”, as it provides information about change over time (or lack thereof) for an outcome of interest aggregated across members of a group. In contrast, “between-individual” stability could be assessed, e.g., by calculating the Spearman correlation between the values obtained at different time points (e.g., ‘disengagement in sixth grade’ with ’disengagement in eighth grade). This analysis focuses on the degree to which individuals retain their relative placement in a group across time. Consider someone who reported the lowest frequencies of disengagement in 6th grade and may report significantly higher disengagement over middle school (i.e., exhibit high levels of change), but report the lowest frequencies of disengagement in eighth grade. That is, the individual is manifesting rank-order stability, even in the context of high mean-level change.

Both types of stability and change are important. Mean-level change in certain traits might help to explain why, in general, populations of individuals tend to be particularly vulnerable to the effects of environmental factors in specific age ranges; rank-order stability might help to quantify the extent to which certain characteristics of the individual are more or less trait-like compared to others. For example, in some areas of development, considerable mean-level change occurs over time (e.g., changes in Big 5 personality traits Bleidorn et al. (2022), but exhibit relatively high rank-order stability, at least over shorter measurement intervals Bleidorn et al. (2022); Roberts and DelVecchio (2000); Roberts, Walton, and Viechtbauer (2006).

Despite the useful information afforded by examining mean-level and rank-order stability and change, these approaches are limited in that they provide little information about the overall patterns of within-individual change and, in turn, can result in fundamental misinterpretations about substantial or meaningful changes in an outcome of interest Patrick J. Curran and Bauer (2011). For example, questions related to the impact of early-onset substance use on brain development focus on changes within a given individual (i.e., intraindividual differences). The ABCD Study will provide researchers with over ten time points for certain constructs (e.g., substance use) across a ten-year period, allowing for a detailed study of some within-person processes.

Stability & Change Diagram.
Figure 1: Types of Stability and Change

3.1.3 Use of appropriate longitudinal models

There is growing recognition that statistical models commonly applied to longitudinal data often fail to align with the developmental theory they are being used to assess (e.g., Patrick J. Curran and Bauer (2011); Hoffman (2015); Andrew K. Littlefield et al. (2021)). First, developmental studies typically involve the use of prospective data to inform theories that are concerned with clear within-person processes (e.g., how phenotypes change or remain stable within individuals over time, (e.g., Patrick J. Curran and Bauer (2011))). Despite this, methods generally unsuited for disaggregating between- and within-person effects (e.g., cross-lagged panel models [CLPM]) remain common within various extant literatures. Fortunately, there exists a range of models that have been proposed to tease apart between- and within-person sources of variance across time (see Andrew K. Littlefield et al. (2021); Orth et al. (2021)). Most of these contemporary alternatives incorporate time-specific latent variables to capture between-person sources of variance and model within-person deviations around an individual’s mean (or trait) level across time (e.g., random-intercept cross-lagged panel model [RI-CLPM], Hamaker, Kuiper, and Grasman (2015)); latent curve models with structured residuals [LCM-SR], Patrick J. Curran et al. (2014)). It is important to note however that these models require multiple assessments waves (e.g., four or more to fully specify the LCM-SR), additional expertise to overcome issues with model convergence, and appreciation of modeling assumptions when attempting to adjudicate among potential models in each research context (see Andrew K. Littlefield et al. (2021), for further discussion).

Second, many statistical models assume certain characteristics about the data to which they are being applied. Common assumptions of parametric statistical models (e.g., linear mixed-effects models) include normality and equality of variances. These assumptions should be carefully considered before finalizing analytical approaches, so that valid inferences can be made from the data, as violation of a model’s assumptions can substantively invalidate the interpretation of results. For example, longitudinal data can exhibit heterogeneous variability (i.e., the variance of the response changes over the duration of the study) that may need to be accounted for within a model. Another pertinent modeling assumption is whether trajectories are linear or non-linear. With two or three assessments per individual, usually only a linear model of within-person change is feasible.

As the study progresses and more time points are assessed, the potentially nonlinear aspects of trajectories can be assessed, for example using quadratic functions of time. Methods that make even fewer assumptions about trajectory shapes, such as nonparametric curve estimation at the mean (e.g., Generalized Additive Mixed Models [GAMMs]; Wood (2017)) and at the individual level (e.g., Functional Data Analysis [FDA]; Ramsay and Silverman (2002)) may also become useful. Note, baseline age in the ABCD Study ranges over two full years; for some outcomes it may be feasible to include a possibly nonlinear effect of baseline age along with a linear effect of within-person change in age even with only two or three assessment times Wesley K. Thompson et al. (2013).

3.1.4 Continuous and Discrete Outcomes

Repeated assessments within the ABCD Study can be based on continuous or discrete measures. Examples of discrete measures include repeated assessments of binary variables (e.g., past 12-month alcohol use disorder status measured across ten years), ordinal variables (e.g., caregiver-reported items measuring emotional and behavioral concerns via the Child Behavior Checklist including the categories of “Not True”, “Somewhat True”, and “Very True”), and count variables (e.g., number of cigarettes smoked per day). In many ways, the distributional assumptions of indicators used in longitudinal designs mirror the decision points and considerations when delineating across different types of discrete outcome variables, a topic that spans entire textbooks (e.g., see Lenz (2016)). For example, the Mplus manual L. K. Muthén (2017) includes examples of a) censored and censored-inflated models, b) linear growth models for binary or ordinal variables, c) linear growth models for a count outcome assuming a Poisson model, and d) linear growth models for a count outcome assuming a zero-inflated Poisson model. Beyond these highlighted examples, other distributions (e.g., negative binomial) can be assumed for the indicators when modeling longitudinal data Ren et al. (2022). These models account for issues that may occur when working with discrete outcomes, including overdispersion, i.e., when the variance is higher than would be expected based on a given parametric distribution (see Lenz (2016)). Given the sheer breadth of issues relevant to determining adequate models for discrete outcomes, it is not uncommon for texts on LDA to only cover models and approaches that assume continuous variables (e.g., T. D. Little (2013)). However, some textbooks on categorical data analysis provide more detailed coverage of the myriad issues and modeling choices to consider when working with discrete outcomes: Lenz (2016), Chapter 11 for matched pair/two-assessment designs; Chapter 12 for marginal and transitional models for repeated designs, such as generalized estimating equations, and Chapter 13 for random effects models for discrete outcomes.

3.1.5 Issues in attributing longitudinal change to development

Systematic changes over time in a variable of interest are not always attributable to development: various pitfalls with longitudinal data can complicate or even invalidate this conclusion. For example, if data missingness or participant dropout are related to the values of the outcome, changing sample composition as the study progresses can bias mean trajectory estimates (we describe this in more detail in Section 3.1.7 below). Another prerequisite for valid developmental interpretations of longitudinal data is to establish whether a construct is measured consistently over time, i.e., longitudinal measurement invariance Liu et al. (2017); Van De Schoot et al. (2015); Willoughby, Wirth, and Blair (2012). Establishing longitudinal measurement invariance ensures that change over time for a given construct is attributable to individual development rather than merely a measurement artifact. For instance, one study using data from the ABCD Study Brislin et al. (2023) found differential item functioning in two items from a brief delinquency measure, revealing significant bias in an arrest item across Black and White youth. More specifically, Black youth were more likely to report being arrested compared to White youth with similar levels of delinquency. Prevalence rates of delinquent behavior would have been severely biased if measurement invariance had not been tested. Alternatively, Vize et al. (2023) showed partially strong to strong evidence of longitudinal measurement invariance across broad externalizing dimensions in youth taking in the ABCD Study, suggesting that changes observed over time in these constructs were not due to systematic measurement error, but likely reflect true developmental change.

Observed patterns of growth and decline often differ between cross-sectional vs. longitudinal effects Salthouse (2014) where subjects gain increasing experience with the assessment with each successive measurement occasion. Such experience effects on cognitive functioning have been demonstrated in adolescent longitudinal samples similar to ABCD Sullivan et al. (2017) and highlight the need to consider these effects and address them analytically. In the case of performance-based measures (e.g., matrix reasoning related to neurocognitive functioning; see Salthouse (2014)), this can be due to “learning” the task from previous test administrations (e.g., someone taking the test a second time performs better than they did the first time simply as a function of having taken it before). Even in the case of non-performance-based measures (e.g., levels of depression), where one cannot easily make the argument that one has acquired some task-specific skill through learning, it has been observed that respondents tend to endorse lower levels on subsequent assessments (e.g., A. T. Beck et al. (1961); French and Sutton (2010)) and this phenomenon has been well documented in research using structured diagnostic interviews Robins (1985). While it is typically assumed that individuals are rescinding or telling us less information on follow-up interviews, there is reason to suspect that in some cases the initial assessment may be artifactually elevated (see Shrout et al. (2018)).

Some longitudinal studies, e.g., accelerated longitudinal designs (ALDs; Wesley K. Thompson et al. (2011)) are especially well suited for discovering these effects and modeling them. While ABCD is not an ALD, the variability in age (and grade in school) at the time of baseline recruitment (approximately 9-10 years old) allows some measures, collected every year, to be conceptualized as an ALD (e.g., substance use; prosocial behavior; family conflict; screen time). It is also possible that in later waves, analyses will allow for disaggregating the confounded effects of age and the number of prior assessments. However, ABCD is fundamentally a single-cohort, longitudinal design, wherein number of prior assessments and age are mostly confounded, and for, perhaps, most analyses, the possible influence of experience effects needs to be kept in mind.

3.1.6 Modeling Covariance

A central issue for repeated measurements on an individual is how to account for the correlated nature of the data. Lack of independence of residuals across time occurs for longitudinal data with repeated assessments on individuals and in other situations with nested data (e.g., visits nested within participants, children nested within schools; siblings nested within families). Note, the ABCD Study has multiple levels of nesting, depending on the analysis, including within-participant, within-family, within-school, within-MRI scanner, and within-site.

Statistical models for nested data include two main components, coupling a model for the mean response and its dependence on covariates with a model for the covariance among repeated outcomes on an individual. In contrast, traditional methods, such as multiple regression and ANOVAs, assume residuals are independent and thus are generally inappropriate for designs that incorporate some type of nesting. Specifically, given that residuals are no longer independent in a repeated measures design, standard errors from these models are biased and can produce misleading inferences. Therefore, an initial question to be addressed by a researcher analyzing prospective data is how to best model their covariance structure. A range of methods can be used to model covariance structures, each with its own set of tradeoffs between model fit and parsimony and which may be more or less appropriate for each specific application (e.g., see Kincaid (2005)).

The most common approach is to use random effects. Essentially, random effects allow for covariance estimates around fixed effects. A classic example (from Bryk and Raudenbush (1992); Singer (1998)) involves math achievement measured among students nested within schools. In a basic, intercept-only model with no covariates (i.e., an unconditional growth model), there would be one fixed effect (the grand mean, or intercept, of math achievement), one school random effect (representing variation in the intercept between schools) and the within-school student residuals (variation left over after accounting for fixed and random effects). In this framework, each student’s score would be the sum of the fixed effect (the grand mean), the school random effect and the student’s within-school residual. Assumptions about the variance and covariance components of this model dictate the form of the variance/covariance structure. For example, if we assume the random effects are independent and identically distributed, the implied structure would be compound symmetry, where it is assumed the covariance of any two students in a single school is captured by a school random intercept and the covariance of any two students in different schools is zero. The assumptions of this relatively simple covariance structure can be relaxed depending on the nesting structure of the data, resulting in different covariance structures with additional parameters (see Singer (1998)).

In longitudinal studies, visits are nested within individuals. Mixed-effect models can be fitted to longitudinal data that couple a model for growth (development) at the mean level with a model for capturing within-individual covariance of assessments. For example, a linear growth model would involve two fixed effects – one for the intercept (the average score when time is coded zero) and one for the linear slope (the change in scores for each unit increase in time). Random effects could include a random effect for intercept, capturing individual variation in scores at time zero, and a random effect for the linear slope, capturing individual variation in linear change across time. Within-individual residuals account for the remaining variation in assessments after accounting for the fixed and random intercepts and slopes. Assumptions regarding the covariation among the random effects also indicate different covariance structures. For example, it is typical to assume that the random intercept and slope components covary, i.e., an individual’s score at time zero relates to the amount of change exhibited across time. Further, particularly in structural equation model forms of this model, it is sometimes assumed that the variance of the residuals varies across assessments Patrick J. Curran (2003).

An alternative to random effects is the autoregressive structure, which allows for correlations between repeated assessments to diminish across time. As the name suggests, the structure assumes the residual of a subsequent measurement occasion (e.g., visit 2) is regressed onto the residual of a prior measurement occasion (e.g., baseline visit). The most common type of autoregressive structure is the AR(1), where residuals at time t + 1 are regressed on residuals at time t. Identical to compound symmetry, this model assumes the variances are homogenous across time; however, it differs from compound symmetry in that the correlations between repeated assessments decline exponentially across visits rather than remaining constant. That is, we can think of the underlying process as a stochastic one that wears itself out over time. For example, per the AR(1) structure, if the correlation between visit 1 and visit 2 data is thought to be .5, then the correlation between visit 1 and visit 3 data would be assumed to be .5 × .5 = .25, and the correlation between visit 1 and visit 4 data would be assumed to be .5 × .5 × .5 = .125. As with compound symmetry, the basic AR(1) model is parsimonious in that it only requires two parameters: the variance of the residuals and the autoregressive coefficient.

Notably, the assumption of constant autoregressive relations between assessments is often relaxed in commonly employed designs that use autoregressive modeling (e.g., CLPM). These designs still typically assume an AR(1) process. However, the magnitude of these relations is often allowed to differ across different AR(1) pairs of assessment (e.g., the relation between visit 1 and visit 2 can be different from the relation between visit 2 and visit 3). These models also often relax the assumption of equal variances of the repeated assessments.

Although the AR(1) structure may involve a more realistic set of assumptions compared to compound symmetry, in that the AR(1) model allows for diminishing correlations across time, the basic AR(1) model, as well as autoregressive models more generally, can also suffer from several limitations in contexts that are common in prospective designs. In particular, recent work demonstrates that if a construct being assessed prospectively across time is trait-like in nature, then a simple AR(1) process fail to adequately account for this trait-like structure, with the downstream consequence that estimates derived from models based on AR structures (such as the CLPM) can be misleading and fail to adequately demarcate between- vs. within-person sources of variance Hamaker, Kuiper, and Grasman (2015). Note also, discrete-time autoregressive structures such as AR(1) implicitly assumes relatively constant time gaps between visits; this may not be true in many applications using the ABCD Study data.

3.1.7 Missing Data/Attrition

Attrition from a longitudinal study such as ABCD is inevitable and represents a potential threat to the external validity of analyses conducted at later visits, especially since attrition can only be expected to grow over time Andrew K. Littlefield et al. (2022). The ABCD Retention Workgroup employs a data-driven approach to examine, track, and intervene in these issues and while preliminary findings show participant race and parent education level to be associated with late and missing visits, although to date, formal attrition in ABCD has been minimal Ewing et al. (2022). Ideally, one tries to minimize attrition through good retention practices from the outset via strategies designed to maintain engagement in the project Cotter et al. (2005); Hill et al. (2016); Watson et al. (2018). However, even the best-executed studies need to anticipate growing attrition over the length of the study and implement analytic strategies designed to provide the most valid inferences.

Perhaps the most key concern when dealing with data that is missing due to attrition is determining the degree of bias in retained variables that is a consequence of attrition. Such bias can attenuate generalizability, particularly if the pattern of missingness is not random (e.g., certain subsets of the population are more likely to drop out/not attend a visit). Assuming that the data are not missing completely at random, attention to the nature of the missingness and employing techniques designed to mitigate attrition-related biases need to be considered in all longitudinal analyses.

Three types of missingness are considered in the literature R. J. Little and Rubin (1989); T. D. Little (2013), namely: a) missing completely at random (MCAR), b) missing at random (MAR), and c) missing not at random (MNAR). Data that are MCAR are a simple random sample of all data in a given dataset. MAR implies missing data are a random sample (i.e., does not hinge on some unmeasured variables) within strata of the measured covariates in a dataset (e.g., biological sex). Data that are MNAR are missing as a function of unobserved variables and may bias associations even after conditioning on the observed covariates. Graham (2009) provides an excellent and easy-to-digest overview of further details involving missing data considerations.

Modern approaches for handling missing data, such as full-information maximum likelihood, propensity weighting, auxiliary variables and multiple imputation avoid the biases of older approaches (see Enders (2010); Graham (2009)). Graham (2009) noted several “myths” regarding missing data. For example, Graham notes many assume the data must be minimally MAR to permit estimating procedures (such as maximum likelihood or multiple imputation) compared to other, more traditional approaches (e.g., using only complete case data). Violations of MAR impact both traditional and more modern data estimation procedures, though as noted by Graham, violations of MAR tend to have a greater effect on older methods. Graham thus suggests that imputing missing data is a better approach compared to listwise deletion in most circumstances, regardless of the model of missingness (i.e., MCAR, MAR, MNAR; see Graham (2009); but also see Twisk et al. (2013)). The ABCD Biostatistics Workgroup is currently implementing several missing data approaches which are being implemented and compared to each other (and listwise deletion) in the 5.0 data release, including, propensity score weighting, and multiple (multilevel) imputation.

3.1.8 Quantifying effect sizes longitudinally

Given that longitudinal data involve multiple sources of variation, quantifying effect sizes longitudinally is more complex compared to deriving such estimates from cross-sectional data. An effect size can be defined as, “a population parameter (estimated in a sample) encapsulating the practical or clinical importance of a phenomenon under study.” (Kraemer 2014). Common effect size metrics include the Pearson correlation r between two variables and the standardized difference between two means, Cohen’s d Cohen (1988). An extensive discussion of cross-sectional effect sizes and their relevance for ABCD is given in Dick et al. (2021).

Adjustments to common effect size calculations, such as Cohen’s d, are required even when only two time points are considered (e.g., Morris and DeShon (2002)). Wang et al. (2019) note there are multiple approaches to obtaining standardized within-person effects, and that commonly suggested approaches (e.g., global standardization) can be problematic (see Wang et al. (2019), for more details). Thus, obtaining effect size metrics based on standardized estimates that are relatively simple in cross-sectional data (such as r) becomes more complex in the context of prospective longitudinal data. Feingold (2009) noted that equations for effects sizes used in studies involving growth modeling analysis (e.g., latent growth curve modeling) were not mathematically equivalent, and the effect sizes were not in the same metric as effect sizes from cross-sectional analysis (see Feingold (2009), for more details).

Given this issue, there have been various proposals for adjusting effect size measures in repeated assessments. Feingold (2019) reviews the approach for effect size metrics for analyses based on growth modeling, including when considering linear and non-linear (e.g., quadratic) growth factors. Morris and DeShon (2002) review various equations for effect size calculations relevant to combining estimates in meta-analysis with repeated measures and independent-groups designs. Other approaches to quantifying effect sizes longitudinally may be based on standardized estimates from models that more optimally disentangle between- and within-person sources of variance. As an example, within a random-intercept cross-lagged panel model (RI-CLPM) framework, standardized estimates between random intercepts (i.e., the correlation between two random intercepts for two different constructs assessed repeatedly) could be used to index the between-person relation, whereas standardized estimates among the structured residuals could be used as informing the effect sizes of within-person relationships.

3.1.9 Longitudinal Data Structures

An ideal longitudinal analysis integrates (a) a well-articulated theoretical model, (b) an appropriate longitudinal data structure, and (c) a statistical model that is an operationalization of the theoretical model Collins (2006). To accommodate various research questions and contexts, different types of longitudinal data and data structures have emerged (see Figure 1). An understanding of these data structures is helpful, as they can warrant different types of LDA. Given that identifying a starting point for making comparisons is somewhat arbitrary, Curran and Bauer and Curran (2019) provide a nice on-ramp in first distinguishing between the use of “time-to-event” and “repeated measures” data. Although both model time, the former is concerned with whether and when an event occurs, whereas the later is focused on growth and change Bauer and Curran (2019) Time-to-event structures measure time from a well-defined origin point up to the occurrence of an event of interest. This data structure is most often analyzed using survival analysis methods (e.g., hazard rate models, event history analysis, failure-time models and the time-to-event data can be based on a single assessment or include multiple recurrent or competing events). While much has been written about “time-to-event” data Hosmer Jr, Lemeshow, and May (2008); Rizopoulos (2012), including a recent analysis examining exclusionary discipline in schools using data from the ABCD Study Brislin et al. (2023), our emphasis will be given to the modeling of “repeated measures” data.

Stability & Change Diagram. Link

When discussing longitudinal analysis, we are most often talking about data collected on the same unit (e.g., individuals) across multiple measurement occasions. However, repeated-measures analysis is not a monolith, and it will serve us well to distinguish between a few of the most common types. One such approach to repeated measures analysis is the use of time-series models. These models generally consist of a long sequence of repeated measurements (≧ 50-100 measurements) on a single or small number of variables of interest. Time-series analysis is often used to predict temporal trends and cyclic patterns and is geared toward making inferences about prospective outcomes within a population (with relatively less focus on inferring individual-level mechanisms and risk factors).

A related type of repeated measures analysis is Intensive Longitudinal Data (ILD). Similar to time-series analysis, ILD models involve frequent measurements (~ 30-40 measurements) of the same individuals in a relatively circumspect period (e.g., experience sampling to obtain time series on many individuals). Although ILD models may include slightly fewer measurement occasions than time-series data, ILD models tend to have more subjects than time-series models (~ 50-100 subjects). This allows ILD models to examine short-term patterns by incorporating a time series model that can sometimes fit parameter estimates to each individual’s data to model individual difference outcomes.

The final type of repeated measures analysis that we will primarily focus on is the longitudinal panel study. These models follow a group of individuals— a panel (also referred to as a cohort) — across relatively fewer measurement occasions (~ 5-15) and are often focused on examining both change within- and between-individuals. The ABCD Study is primarily a longitudinal panel study, though some data streams (e.g., functional brain imaging, FitBit data) could be analyzed as ILP or even time series methods.

While other longitudinal designs have their own unique strengths and applications, the longitudinal panel design is particularly well-suited for investigating developmental processes in the context of the ABCD Study. In the following sections, we will discuss various analytic methods commonly used to analyze longitudinal panel data, including growth models, mixed models, and a number of additional trajectory models. These methods provide valuable insights into within- and between-individual differences and are highly relevant for researchers working with the ABCD Study dataset. By focusing on these methods, we aim to equip readers with the knowledge necessary to conduct longitudinal research and perform analyses using the rich, longitudinal, and publicly available data from the ABCD Study.

4 Longitudinal Analysis

4.0.1 Types of longitudinal panel models

With the large and continually expanding body of research on statistical methods for longitudinal analyses, determining which longitudinal model to implement can be challenging. This section aims to help researchers navigate these many options to identify the statistical approach most appropriate to their unique research question when deciding on how to measure change over time. Notably, there are a myriad of viable ways one can go about grouping various types of longitudinal models for presentation.

Common examples include grouping by linear vs nonlinear models Collins (2006), the number of measurement occasions King et al. (2018), and statistical equivalency (e.g., change scores vs. residualized change; see Castro-Schilo and Grimm (2018)). The organization we use below overlaps in several ways with these examples, and in particular with Bauer and Curran (2019). However, it is important to note that in each case, the chosen way of grouping is primarily intended to allow the reader to compare and contrast various analytical approaches. In the following sections, we briefly summarize the advantages/disadvantages of a series of longitudinal models organized into the following groupings: Traditional Models, Modern GLM Extensions, Structural Equation Models (SEM), and Advanced SEM (see Figure 2). We note that this is not an exhaustive review of each of these methods, and for more in-depth detail we do provide the reader with relevant resources. As aptly summarized by Bauer and Curran (2019), “…there are many exceptions, alternatives, nuances, ‘what ifs’, and ’but couldn’t you’s that aren’t addressed here.”

Longitudinal Models & Data Structures
Figure 2: Longitudinal Models/Data Structures

Traditional Models

Traditional methods for longitudinal analysis primarily focus on modeling mean-level change, and how these changes may differ across groups or levels of some other variable. For example, is there a difference in average internalizing symptoms obtained across multiple assessments between boys and girls? Longitudinal models that focus on mean-level change are also referred to as marginal models and examples of specific methods include repeated measures ANOVA, ANCOVA and Generalized Estimating Equations (GEEs). Mean-level change models are commonly used when data are only available from 2 measurement occasions. For example, computing a difference score (e.g., mean internalizing scores at visit 2 - mean internalizing scores at visit 1) that can be used as an outcome in a subsequent GLM analysis (e.g., paired-samples t-test, repeated measures ANOVA) to test for differences in patterns of change over time and between groups. Additionally, the longitudinal signed-rank test, a nonparametric alternative to the paired t-test, can be a useful tool for analyzing non-normal paired data. Another common approach, often used in pre-/post-design studies but can be used with ABCD Study data, is to use residualized change score analysis to assess the degree of change in a variable, while controlling for its initial level Castro-Schilo and Grimm (2018).

For example, to examine change in cortico-limbic connectivity among ABCD participants, Brieant, Sisk, and Gee (2021) regressed cortico-limbic connectivity at the year 2 follow-up on baseline cortico-limbic connectivity, which allowed the authors to examine the associations between negative life events and the variance of cortico-limbic connectivity unexplained by baseline connectivity. Similarly, Romer and Pizzagalli (2021) used a residualized-change model to examine the bidirectional influences of executive functioning and a general psychopathology factor ‘p’ across the first two years of the ABCD Study. Both studies were able to conclude associations between their constructs of interest that could not be accounted for by prior frequencies at baseline.

Traditional longitudinal models, such as residualized change score models, can be useful in some contexts (e.g., two measurement occasions), but overall, their practical utility for answering questions about developmental processes is limited. Perhaps most notably, these models do not allow for characterizing patterns of within-person change. This is a particularly important limitation since most psychological theories posit within-person processes (i.e., what will happen within a given individual). As such, traditional approaches often correspond poorly with most theoretical models of change and a failure to disaggregate between-person and within-person effects can result in consequential errors of inference (e.g., ecological fallacy, Patrick J. Curran and Bauer (2011)). Moreover, even determining which of these procedures to use for comparing change over two time points across groups can be surprisingly complicated. A particularly vexing example is that of imbalanced baseline scores (i.e., when baseline scores are correlated with a covariate of interest), which can produce different conclusions across methods (e.g., see Andrew K. Littlefield (2023), for a review). Given these shortcomings, and the complexity of the issues surrounding some of these methods, it is typically recommended that researchers make use of more modern approaches for analyzing longitudinal data and preferably make use of data collected across three or more time points, as is currently true for many ABCD Study assessments.

Modern GLM Extensions

Modern approaches to LDA have advanced beyond traditional methods by offering greater flexibility and a more in-depth understanding of within-person and between-person variability. Generalized Estimating Equations (GEE), Linear Mixed Models (LMM), Generalized Linear Mixed Models (GLMM), and Autoregressive Cross-Lagged Panel Models (ARCL) are examples of such contemporary techniques. GEE, an extension of Generalized Linear Models, combines the generalized linear model for non-normal outcomes with repeated measures and is suitable for analyzing correlated longitudinal data and modeling population-averaged effects. For example, Van Dijk et al. (2021) used GEE to obtain relative risks for psychiatric diagnoses among children in the ABCD Study with a family history of depression and used the ABCD Study sampling weights to generalize prevalence rates among 9 and 10-year-olds across the US.

LMMs, also known as multilevel or hierarchical linear models, facilitate the simultaneous analysis of within-person and between-person variability, making them ideal for nested data structures or repeated measures. Within the ABCD Study, researchers may want to consider nesting by individual, family (i.e., siblings or twins), school or district, and/or site. GLMMs further extend the LMM framework to accommodate non-normal response variables, such as binary, count, or ordinal data, such as the use of ABCD data on substance use (e.g., Martz et al. (2022)) screen media use Lees et al. (2020), and microstructure of the brain Palmer et al. (2022).

Finally, ARCL models are used to investigate reciprocal relationships between variables over time, as they estimate both autoregressive and cross-lagged effects, although ARCL models are relatively less useful for teasing apart between-person and within-person sources of variances; see Patrick J. Curran and Hancock (2021).

The strengths of these modern methods lie in their ability to account for individual differences, within-person change, and time-varying predictors, thereby providing a more comprehensive understanding of complex relationships in longitudinal data. Despite these advantages, modern approaches may require more complex modeling assumptions and higher computational demands compared to traditional methods. Additionally, proper model specification and the interpretation of results can be more challenging, especially in cases of high multicollinearity or missing data. However, modern longitudinal analysis methods have generally surpassed traditional methods in addressing a wider range of research questions, accommodating diverse data structures, and elucidating the intricate dynamics of developmental processes.

Structural Equation Modeling (SEM)

Structural Equation Modeling (SEM) is a flexible modeling framework that integrates elements of path analysis and confirmatory factor analysis (CFA) to examine complex relationships between a set of observable variables and latent constructs Hair Jr et al. (2021). The integration of structural (regression) and measurement (CFA) components within a unified framework supports a theory-driven approach that allows researchers to rigorously test hypothesized relationships among variables of interest and their underlying causes Hair Jr et al. (2021); Raykov and Marcoulides (2012). Over the years, the flexibility of the SEM framework has evolved to become particularly adept for modeling autoregressive processes (which often assume underlying stationarity) and growth processes which accommodate both, mean trajectories and individual differences in them J. J. McArdle (2009); T. D. Little (2013).

Longitudinal SEM techniques share many similarities with mixed-effects methods and research demonstrates their mathematical equivalence in many situations Patrick J. Curran (2003); Mehta and Neale (2005). However, these related approaches often cater to distinct theoretical and analytical needs. For instance, mixed-effects techniques are an extension of the regression framework and often excel when working with complex data structures such as multiple levels of nesting, small samples, and non-equidistant time points McNeish and Matta (2018); Hedeker and Gibbons (2006). Alternatively, applying SEM methods to longitudinal analysis provides a flexible means for modeling the underlying process of change. It also addresses several challenges faced by competing approaches, including the ability to accommodate intricate error structures and deal effectively with missing data, as well as the implementation of numerous modeling extensions McNeish and Matta (2018); Patrick J. Curran and Hancock (2021). These models have grown increasingly popular for modeling longitudinal outcomes particularly due to their ability to build statistical models that match some particular underlying theory Serang, Grimm, and Zhang (2019).

Considering the variety of available techniques, it can be helpful to classify longitudinal SEM, broadly (if not coarsely), into variable-centered, person-centered, and hybrid analyses, each with unique strengths and limitations. Variable-centered analyses (e.g., latent growth curves Patrick J. Curran (2003), latent change scores J. J. McArdle and Hamagami (2001), latent state-trait models Geiser and Lockhart (2012)) are primarily concerned with understanding covariation among variables at the group level and characterizing population-level patterns of change, while person-centered analyses (e.g., latent class and latent transition models) identify distinct subgroups or patterns within the data B. Muthén and Muthén (2000); Woo et al. (2024); Howard and Hoffman (2018). Hybrid models combine these perspectives to offer a comprehensive analysis of latent subgroups and growth parameter relationships Morin, Bujacz, and Gagné (2018); G. H. Lubke and Muthén (2005). The choice between these approaches is primarily driven by the research question, data structure, and relevant underlying assumptions.

Variable-centered models

One key application of the SEM framework to the analysis of longitudinal data is the latent growth curve model (LGCM). This is a variable-centered approach that characterizes average group trajectories and individual variations (random effects) in an outcome over time Patrick J. Curran (2003). These models are similar to their linear mixed effects counterpart in many ways, with the main conceptual difference being that LGCM includes a repeatedly measured outcome in the model as a function of time (closely resembling a standard CFA approach), rather than as an explanatory variable (as in a standard regression approach) McNeish and Matta (2018). Specifically, observed scores at each time point are treated as indicator variables with their factors loading scaled to reflect a hypothesized pattern of change (e.g., loadings of 0, 1, and 2 would assume equidistant, linear change). Latent intercepts (initial levels) and slopes (rates of change) are estimated, along with their variances and covariance to capture common trends and individual deviations over time. This method was used in a recent study by Trevino et al. (2023) to show a decreasing trajectory of parent-reported externalizing behaviors from ages 9-12 among youth taking part in the ABCD Study. This study also examined hypothesized predictors of the growth trajectory intercept and slope factors, highlighting a particular strength of these models–– their flexibility and extensibility. As an example, Roy et al. (2024) used publicly available data from the ABCD Study and several other large-scale datasets to explore bivariate (parallel process) relationships between white matter pathways and literacy over time. Beyond these examples, LGCMs can be extended in numerous ways, including to compare rates of growth across groups, investigate the consequences of change, and incorporate time-invariant or time-varying covariates, to highlight only a few (for a more detailed treatment of LGCM applications and methodologies, refer to Preacher, Zyphur, and Zhang (2010); Preacher (2018); Patrick J. Curran, Obeidat, and Losardo (2010)).

The latent change score model (LCSM) is a variable-centered approach uniquely tailored for analyzing temporal variations in how a construct changes over time J. McArdle and Nesselroade (1994); J. J. McArdle and Hamagami (2001). These models share many features with growth curve analysis, but with a more explicit focus on how change occurs between measurement occasions Serang, Grimm, and Zhang (2019); J. J. McArdle (2009). Specifically, LCSM estimates a series of latent variables to model change in an outcome from one time point to the next, as a function of scores on that outcome at prior time points J. J. McArdle and Hamagami (2001); Ghisletta and McArdle (2012). Some types of LCSM estimate two underlying latent factors: a constant change factor that remains fixed over time, and a proportional change factor that adjusts for previous scores. By disaggregating change into constant and proportional components, this approach facilitates a more nuanced understanding of whether prior changes in a given process are related to future changes in the same process Serang, Grimm, and Zhang (2019); Kievit et al. (2018). Expanding upon the capabilities of this framework, LCSM also allows for comprehensive multivariate analyses that can facilitate investigations into how change in one construct is associated with change in another construct. The appeal of this approach is evidenced by several recent studies that have used data from the ABCD Study to explore bivariate associations between brain development and changes in several mental and physical health indicators Wiker et al. (2023); Rapuano et al. (2022); D. Beck et al. (2023); Nweze et al. (2023); Mewton et al. (2023).

Latent State-Trait Models (LSTM) offer another variable-centered approach to longitudinal analysis that also allows for the estimation of patterns of change over time. Unlike LGCM, which conceptualizes change as a function of time, and LCSM, which views change through sequential measurements, this approach disaggregates observed behaviors into distinct stable (trait) and occasion-specific (state) components Kenny and Zautra (2001); Steyer et al. (2015). Based on LST theory Steyer, Schmitt, and Eid (1999); Steyer, Ferring, and Schmitt (1992), these models hold that scores on a repeated measures outcome can be partitioned into an enduring latent trait variable that reflects between-individual differences, and a transient latent state residual that represents situational influences Stadtbaeumer, Kreissl, and Mayer (2022); Geiser and Lockhart (2012). Beyond parsing out these key variance components, LSTM can be extended in many ways, such as by incorporating autoregressive effects to capture relative stability and the influence of past states on future responses (i.e. carry-over effects; Cole, Martin, and Steiger (2005); Eid et al. (2017); Geiser and Lockhart (2012)). The merits of this approach are highlighted in a recent review by Sanchez-Alonso and Aslin (2020) focused on strategies for modeling neurobehavioral development. These study authors encourage researchers to leverage data from the ABCD Study and other large-scale longitudinal and publicly available datasets and to apply state-trait methods to map neural and behavioral trajectories in youth (for a more detailed overview of these models, see Kenny and Zautra (2001); Steyer, Geiser, and Loßnitzer (2023); Steyer, Schmitt, and Eid (1999)). In general, while many commonalities and important features are shared across different variable-centered approaches, selecting the most appropriate statistical model for assessing change hinges on the specific theoretical model of change and what is intended to be learned from the model (see Kievit et al. (2018); J. J. McArdle (2009); Ghisletta and McArdle (2012), for discussion), which is critical for informing the interpretation and applicability of the research findings.

Person-centered models

Despite the flexibility afforded by variable-centered analysis, these methods are not generally equipped to capture underlying developmental trajectories that are unique to distinct clusters of individuals. This limitation can be particularly notable for research that aims to characterize heterogeneous developmental processes. Person-centered approaches, including latent transition analysis and latent class growth analysis, address this limitation by identifying subgroups of individuals who share similar patterns of change. These models can reveal meaningful subpopulations and help researchers understand the factors that contribute to differences in developmental trajectories. For example, taking advantage of the large sample size of the ABCD Study, Xiang et al. (2022) found evidence of four subgroups of youth with unique longitudinal patterns of depressive symptoms over time and identified risk factors that were differentially associated with the various trajectories.

The use of such models allows for a more nuanced understanding of the associations between risk factors and change in symptomatology, as opposed to a snapshot of symptomatology at a single time point. Despite a range of potential model specifications for longitudinal mixture modeling, person-centered approaches tend to use parameterizations that default to settings found in popular software packages (e.g., Mplus). It has recently been demonstrated (see McNeish and Harring (2021)) that the use of such specifications tends to identify the so-called “cat’s cradle” solution (see Sher, Jackson, and Steinley (2011)) that consists of “…(a) a consistently ‘low’ group, (b) an ‘increase’ group, (c) a ‘decrease’ group, and (d) a consistently ‘high’ group” Sher, Jackson, and Steinley (2011), p. 322. Indeed, Xiang et al. (2022) describe their four-group solution as follows: “Of all participants, 536 (10.80%) were classified as increasing, 269 (5.42%) as persistently high, 433 (8.73%) as decreasing, and 3724 (75.05%) as persistently low” Xiang et al. (2022), p. 162. Although Sher, Jackson, and Steinley (2011) cautioned that groups from these trajectory-based approaches should not be over-reified, this practice also remains common (e.g., Hawes et al. (2016); Hawes et al. (2018)). Thus, though person-centered approaches can, in theory, help researchers understand the factors that contribute to differences in developmental trajectories, researchers should more thoughtfully consider alternative specifications (see Andrew K. Littlefield, Sher, and Steinley (2010), as an example) and be especially skeptical when default specifications identify these four prototypic groups.

Hybrid approaches, such as growth mixture (jung2008; B. Muthén and Muthén (2000)) and factor-mixture G. H. Lubke and Muthén (2005); G. Lubke and Muthén (2007) modeling, combine aspects of both variable-centered and person-centered models, allowing for the identification of latent subgroups while also modeling relationships among growth parameters. This combination provides a more comprehensive understanding of longitudinal data by capturing both within- and between-person variability. However, hybrid models can be more complex, necessitating careful model specification, selection, and interpretation. Additionally, these methods may require larger sample sizes to ensure the stability and accuracy of results.

In summary, SEM approaches offer powerful tools for LDA, enabling researchers to investigate complex relationships, individual differences, and change dynamics over time. The choice between variable-centered, person-centered, and hybrid approaches depends on the research objectives and the nature of the data. Despite their limitations, these models have greatly advanced our understanding of developmental processes and the factors that contribute to individual differences in change trajectories.

Advanced Structural Equation Models

Advanced SEM approaches, such as the RI-CLPM and LCM-SR models, have emerged to provide a clearer understanding of important research questions and data structures in longitudinal analysis. These advanced models extend traditional SEM techniques, enabling researchers to disentangle within-person and between-person effects, as well as capture additional time-specific dependencies and associations that may not be accounted for by the latent growth factors.

The RI-CLPM enhances the traditional cross-lagged panel model by incorporating random intercepts, which allow for the separation of stable individual differences from the dynamic within-person associations between variables over time. Within-person variance in these models is captured by a series of latent variables that reflect time specific variance (i.e., the residual variance from the random intercept). These time-specific variables are referred to as structured residuals. Distinguishing between-person variance subsumed by the random intercept from the structured residuals is particularly valuable for understanding the time-specific effects of one variable on another, while accounting for the influence of individual differences. However, RI-CLPM may require larger sample sizes to ensure stability and accuracy of the estimates and can be computationally demanding. Using three waves of ABCD Study data, Kulisch et al. (2023) found a prospective association between psychopathology and childhood obesity as well as between childhood obesity and later eating behavior. The authors also showed that reciprocal associations were overestimated when stable, interindividual trait differences were not included in the model (i.e., via the random intercept).

LCM-SR, on the other hand, extends the RI-CLPM by including additional growth factors, such as a random linear slope. That is, the LCM-SR is a hybrid between a latent growth model and CLPM. This approach allows for a more comprehensive understanding of within-person change dynamics and factors influencing change over time. By including structured residuals, LCM-SR can capture additional time-specific relationships that are not explained by the latent growth factors. However, even more so than the RI-CLPM, LCM-SR comes with increased model complexity and requires careful specification and interpretation.

In conclusion, advanced SEM approaches for LDA provide valuable tools for addressing complex research questions and data structures. While they offer more nuanced insights into within-person change dynamics and the influence of individual differences, these models also come with certain limitations, such as the necessity of multiple assessments (e.g., four or more for LCM-SR), increased complexity, computational demands, and the need for careful model specification and interpretation. As with any statistical method, researchers should carefully consider their research objectives, data characteristics, and the assumptions of each model when selecting the most appropriate advanced SEM approach for longitudinal analysis. Given that these modeling approaches necessitate more waves of data, they are not yet commonly used with ABCD Study data. We anticipate that as more waves of ABCD data are publically released, these models can be used to address some of the pitfalls of the more traditional methods.

Longitudinal Analysis of Neuroimaging Data

Neuroimaging data, characterized by its large scale, spatial structure and binary data formats, requires the use of specialized software for effective analysis. Fortunately, there are now several freely available software packages that provide options for statistical modeling of brain imaging data, thus facilitating analysis of the brain’s function or structure at every voxel or vertex in an image (see Fig. 3). However, the most widely used packages have only rudimentary support for longitudinal data. Prominent software solutions such as SPM (https://www.fil.ion.ucl.ac.uk/spm) and FSL (https://fsl.fmrib.ox.ac.uk) packages offer limited support in analyzing longitudinal data due to their use of strong assumptions. SPM assumes a common longitudinal correlation structure over space, and FSL requires balanced designs and relies on the assumption of compound symmetry. Despite these limitations, there is a steady growth in the development of neuroimaging tools that provide for comprehensive longitudinal data analysis, typically via commonly adopted modeling approaches such as LMMs marginal models. These emerging tools are equipped to handle more complex scenarios, including unbalanced designs and random covariate effects, among others.

AFNI (https://afni.nimh.nih.gov/), a well-established tool in neuroimaging, integrates 3dLMEr Chen et al. (2013); https://afni.nimh.nih.gov/pub/dist/doc/program_help/3dLMEr.html which adopts an LMM-based approach by providing access to the advanced capabilities of R’s lme4’s lmer function. For surface-based data, Freesurfer (https://surfer.nmr.mgh.harvard.edu/) provides the linear mixed effects (LME) package for modelling longitudinal data Bernal-Rusiel, Greve, et al. (2013); https://surfer.nmr.mgh.harvard.edu/fswiki/LinearMixedEffectsModels. This tool is able to apply spatial regularization of LMM parameters with surface-based ROIs to improve stability Bernal-Rusiel, Reuter, et al. (2013).

For imaging data, LMM’s present a significant computational challenge, not only because they require iterative optimization but also because the computations cannot be vectorized as efficiently compared to ordinary least squares. To overcome this challenge, The Big Linear Mixed Models (BLMM, https://github.com/NISOx-BDI/BLMM) software addresses this by using Python’s broadcasting operations to estimate LMM’s as efficiently as possible Maullin-Sapey and Nichols (2022); Maullin-Sapey and Nichols (2021). BLMM further separates the computation of sufficient statistics and parameter estimation, allowing sensitive image data to remain private if needed.

A different yet efficient approach is used with Fast and Efficient Mixed-effects Analysis (FEMA, https://github.com/cmig-research-group/cmig_tools), which uses a non-iterative regression estimator of the LMM variance components plus variance parameter quantization. This allows vectorization within groups of voxels that share the same variance parameters Parekh et al. (2021). While this method uses different approximations, the authors have shown it provides results that closely match a traditional LMM implementation.

An alternative method for modeling longitudinal data is the marginal model. This approach differs from others by modeling only the population-level factors and covariates rather than explicitly modeling individual intercepts and slopes. It employs a flexible intra-subject covariance model to account for residual dependence. Like the LMM, it allows for unbalanced designs and singleton subjects, and it implicitly measures any covariance that would otherwise be explained by random covariates. In a marginal model, a “working covariance” matrix is utilized. This matrix does not need to be accurately modelled and may even be constructed under the potentially incorrect assumption that the errors are independent. In this case, parameter estimation reduces to ordinary least squares, but remains consistent. A crucial component of this method is the use of a robust “sandwich estimator” for standard errors, which accounts for any dependence not captured by the working covariance. The Sandwich Estimator (SwE) is a toolbox for SPM (https://www.nisox.org/Software/SwE, including CIFTI support) and FSL (https://fsl.fmrib.ox.ac.uk/fsl/fslwiki/Swe) that provides marginal model inference using an independence working covariance matrix Guillaume et al. (2014).

Statistical modeling of brain imaging data Statistical modeling of brain imaging data

5 Discussion

As we enter the era of large-scale longitudinal investigations, it is essential to critically examine the various analytical methods that can be employed to glean insights from these rich datasets. The complex nature of longitudinal data demands sophisticated and well-suited methodologies to accurately address research questions and minimize biases. This paper aimed to provide an overview of diverse longitudinal analysis techniques, with a particular emphasis on their application to extensive longitudinal studies such as the ABCD Study. Beyond contributing to the ever-growing body of knowledge on LDA, we hope this manuscript also serves as a valuable resource for researchers seeking to optimize the use of large-scale longitudinal investigations in advancing our understanding of human development and behavior. In this discussion, we will focus on the key findings and recommendations of our review and discuss potential innovations that can further enhance the utility of these methods.

We began by addressing fundamental concepts and considerations in longitudinal research that are essential for generating accurate and meaningful insights into developmental processes. Concepts such as vulnerable periods, developmental disturbances and snares, or cascade and experience effects (among many others), are instrumental in shaping the design, analysis, and interpretation of longitudinal studies. Together, these concepts provide a framework for understanding the mechanisms underlying the course of development, while also accounting for the complex interplay between individual development and the influence of environmental factors. By considering the intricate relationships among these factors, researchers can better identify the critical time periods, situations, and contexts that contribute to individual differences in developmental outcomes. This awareness enables more precise inferences regarding the causal relationships between exposures and outcomes, ultimately leading to more robust and meaningful findings that can help facilitate the translation of research findings into practical applications in clinical and public health settings.

We also discussed some of the opportunities, challenges, and pitfalls that arise when working with longitudinal data. Key issues include selecting appropriate methods to account for the intricacies of longitudinal data, addressing missing data in a way that minimizes biases, and determining suitable longitudinal data structures that align with research questions and context. To address these challenges, researchers should carefully consider issues such as study design, selection of methods that account for both within- and between-person sources of variance, and employing modern techniques, (e.g., FIML, multiple imputation) for handling missing data. By adhering to best practices in longitudinal research and remaining vigilant of potential pitfalls, researchers can effectively harness the power of longitudinal data to maximize the potential of their investigations and gain valuable insights into complex developmental processes, individual differences, and the underlying mechanisms that drive change over time.

The final section, along with associated code and additional resources made available as online supplements, aims to serve as a resource for researchers seeking to understand and implement various longitudinal panel models. By providing an overview of different approaches, their strengths and limitations, and key considerations for their use, we hope to facilitate the selection of appropriate models tailored to specific research questions and data structures. It is essential for researchers to consider their research objectives, the characteristics of their data, and the assumptions underlying each model when choosing the most suitable approach for longitudinal analysis.

We encourage researchers to consult the cited literature and online supplements for further guidance in selecting and implementing longitudinal models when using the ABCD Study dataset. As the field continues to advance, we anticipate the emergence of new methods and refinements to existing approaches, further expanding the toolkit available to researchers for the analysis of longitudinal data. By staying informed about developments in this area and critically evaluating the appropriateness of different models for their research questions, researchers can ensure that their longitudinal analyses are both rigorous and informative. Notably, in this vast and continually evolving field, with numerous models and approaches available to address a wide range of research questions, no single model is universally applicable or without limitations. The diversity of methods ensures that researchers can find an appropriate tool for their specific needs. By familiarizing themselves with the various types of longitudinal models, researchers can more effectively navigate the complexities of longitudinal data and contribute valuable insights into the developmental processes and individual differences that shape human experience.


6 Acknowledgements

Data used in the preparation of this article were obtained from the Adolescent Brain Cognitive DevelopmentSM (ABCD) Study (https://abcdstudy.org), held in the NIMH Data Archive (NDA). This is a multisite, longitudinal study designed to recruit more than 10,000 children age 9-10 and follow them over 10 years into early adulthood. The ABCD Study® is supported by the National Institutes of Health and additional federal partners under award numbers U01DA041048, U01DA050989, U01DA051016, U01DA041022, U01DA051018, U01DA051037, U01DA050987, U01DA041174, U01DA041106, U01DA041117, U01DA041028, U01DA041134, U01DA050988, U01DA051039, U01DA041156, U01DA041025, U01DA041120, U01DA051038, U01DA041148, U01DA041093, U01DA041089, U24DA041123, U24DA041147. A full list of supporters is available at https://abcdstudy.org/federal-partners.html. A listing of participating sites and a complete listing of the study investigators can be found at https://abcdstudy.org/consortium_members/. ABCD consortium investigators designed and implemented the study and/or provided data but did not necessarily participate in the analysis or writing of this report. This manuscript reflects the views of the authors and may not reflect the opinions or views of the NIH or ABCD consortium investigators.

The ABCD data repository grows and changes over time. The ABCD data used in this report came from DOI: 10.15154/z563-zd24. DOIs can be found at the following link.

References

Arseneault, Louise, Mary Cannon, Richie Poulton, Robin Murray, Avshalom Caspi, and Terrie E Moffitt. 2002. “Cannabis Use in Adolescence and Risk for Adult Psychosis: Longitudinal Prospective Study.” Bmj 325 (7374): 1212–13.
Bagot, K. S., S. A. Matthews, M. Mason, Lindsay M. Squeglia, J. Fowler, K. Gray, M. Herting, A. May, Ian Colrain, and J. Godino. 2018. “Current, Future and Potential Use of Mobile and Wearable Technologies and Social Media Data in the ABCD Study to Increase Understanding of Contributors to Child Health.” Developmental Cognitive Neuroscience 32: 121–29.
Barch, Deanna M., Matthew D. Albaugh, Shelli Avenevoli, Linda Chang, Duncan B. Clark, Meyer D. Glantz, James J. Hudziak, Terry L. Jernigan, Susan F. Tapert, and Debbie Yurgelun-Todd. 2018. “Demographic, Physical and Mental Health Assessments in the Adolescent Brain and Cognitive Development Study: Rationale and Description.” Developmental Cognitive Neuroscience 32: 55–66.
Baron, Reuben M, and David A Kenny. 1986. “The Moderator–Mediator Variable Distinction in Social Psychological Research: Conceptual, Strategic, and Statistical Considerations.” Journal of Personality and Social Psychology 51 (6): 1173.
Bauer, D. J., and P. J. Curran. 2019. “Conducting Longitudinal Data Analysis: Knowing What to Do and Learning How to Do It.” https://centerstat.org/wp-content/uploads/2019/03/SRCD2019-2up.pdf.
Bechtold, Jordan, Alison Hipwell, David A Lewis, Rolf Loeber, and Dustin Pardini. 2016. “Concurrent and Sustained Cumulative Effects of Adolescent Marijuana Use on Subclinical Psychotic Symptoms.” American Journal of Psychiatry 173 (8): 781–89.
Beck, Aaron T., Calvin H. Ward, Mock Mendelson, Jeremiah Mock, and John Erbaugh. 1961. “An Inventory for Measuring Depression.” Archives of General Psychiatry 4 (6): 561–71.
Beck, Dani, Lia Ferschmann, Niamh MacSweeney, Linn B Norbom, Thea Wiker, Eira Aksnes, Valerie Karl, et al. 2023. “Puberty Differentially Predicts Brain Maturation in Male and Female Youth: A Longitudinal ABCD Study.” Developmental Cognitive Neuroscience, 101261.
Bernal-Rusiel, Jorge L, Douglas N Greve, Martin Reuter, Bruce Fischl, Mert R Sabuncu, Alzheimer’s Disease Neuroimaging Initiative, et al. 2013. “Statistical Analysis of Longitudinal Neuroimage Data with Linear Mixed Effects Models.” Neuroimage 66: 249–60.
Bernal-Rusiel, Jorge L, Martin Reuter, Douglas N Greve, Bruce Fischl, Mert R Sabuncu, Alzheimer’s Disease Neuroimaging Initiative, et al. 2013. “Spatiotemporal Linear Mixed Effects Modeling for the Mass-Univariate Analysis of Longitudinal Neuroimage Data.” Neuroimage 81: 358–70.
Bind, M-AC, TJ Vanderweele, BA Coull, and JD Schwartz. 2016. “Causal Mediation Analysis for Longitudinal Data with Exogenous Exposure.” Biostatistics 17 (1): 122–34.
Bleidorn, Wiebke, Ted Schwaba, Anqing Zheng, Christopher J Hopwood, Susana S Sosa, Brent W Roberts, and DA Briley. 2022. “Personality Stability and Change: A Meta-Analysis of Longitudinal Studies.” Psychological Bulletin 148 (7-8): 588.
Brieant, Alexis E, Lucinda M Sisk, and Dylan G Gee. 2021. “Associations Among Negative Life Events, Changes in Cortico-Limbic Connectivity, and Psychopathology in the ABCD Study.” Developmental Cognitive Neuroscience 52: 101022.
Brislin, Sarah J, D Angus Clark, Duncan B Clark, C Emily Durbin, Ashley C Parr, Lia Ahonen, Kaston D Anderson-Carpenter, et al. 2023. “Differential Item Functioning in Reports of Delinquent Behavior Between Black and White Youth: Evidence of Measurement Bias in Self-Reports of Arrest in the Adolescent Brain Cognitive Development Study.” Assessment, 10731911231164627.
Bryk, Anthony S, and Stephen W Raudenbush. 1992. Hierarchical Linear Models: Applications and Data Analysis Methods. Sage Publications, Inc.
Casey, B. J., Tariq Cannonier, May I. Conley, Alexandra O. Cohen, Deanna M. Barch, Mary M. Heitzeg, Mary E. Soules, et al. 2018. “The Adolescent Brain Cognitive Development (ABCD) Study: Imaging Acquisition Across 21 Sites.” Developmental Cognitive Neuroscience, The Adolescent Brain Cognitive Development (ABCD) Consortium: Rationale, Aims, and Assessment Strategy, 32 (August): 43–54. https://doi.org/10.1016/j.dcn.2018.03.001.
Castro-Schilo, Laura, and Kevin J. Grimm. 2018. “Using Residualized Change Versus Difference Scores for Longitudinal Research.” Journal of Social and Personal Relationships 35 (1): 32–58. https://doi.org/10.1177/0265407517718387.
Chen, Gang, Ziad S Saad, Jennifer C Britton, Daniel S Pine, and Robert W Cox. 2013. “Linear Mixed-Effects Modeling Approach to FMRI Group Analysis.” Neuroimage 73: 176–90.
Cohen, Jacob. 1988. “Statistical Power.” Analysis for the Behavioral Sciences, 273–406.
Cole, David A, Nina C Martin, and James H Steiger. 2005. “Empirical and Conceptual Problems with Longitudinal Trait-State Models: Introducing a Trait-State-Occasion Model.” Psychological Methods 10 (1): 3.
Collins, Linda M. 2006. “Analysis of Longitudinal Data: The Integration of Theoretical Model, Temporal Design, and Statistical Model.” Annu. Rev. Psychol. 57: 505–28.
Cotter, Robert B., Jeffrey D. Burke, Magda Stouthamer-Loeber, and Rolf Loeber. 2005. “Contacting Participants for Follow-up: How Much Effort Is Required to Retain Participants in Longitudinal Studies?” Evaluation and Program Planning 28 (1): 15–21.
Crews, Fulton T, Christopher J Braun, Blair Hoplight, Robert C Switzer III, and Darin J Knapp. 2000. “Binge Ethanol Consumption Causes Differential Brain Damage in Young Adolescent Rats Compared with Adult Rats.” Alcoholism: Clinical and Experimental Research 24 (11): 1712–23.
Curran, Patrick J. 2003. “Have Multilevel Models Been Structural Equation Models All Along?” Multivariate Behavioral Research 38 (4): 529–69.
Curran, Patrick J., and Daniel J. Bauer. 2011. “The Disaggregation of Within-Person and Between-Person Effects in Longitudinal Models of Change.” Annual Review of Psychology 62: 583–619.
Curran, Patrick J, and Gregory R Hancock. 2021. “The Challenge of Modeling Co-Developmental Processes over Time.” Child Development Perspectives 15 (2): 67–75.
Curran, Patrick J., Andrea L. Howard, Sierra Bainter, Stephanie T. Lane, and James S. McGinley. 2014. “The Separation of Between-Person and Within-Person Components of Individual Change Over Time: A Latent Curve Model with Structured Residuals.” J Consult Clin Psychol 82 (5): 879–94. https://doi.org/10.1037/a0035297.
Curran, Patrick J, Khawla Obeidat, and Diane Losardo. 2010. “Twelve Frequently Asked Questions about Growth Curve Modeling.” Journal of Cognition and Development 11 (2): 121–36.
Dick, Anthony Steven, Daniel A Lopez, Ashley L Watts, Steven Heeringa, Chase Reuter, Hauke Bartsch, Chun Chieh Fan, et al. 2021. “Meaningful Associations in the Adolescent Brain Cognitive Development Study.” NeuroImage 239: 118262.
Duncan, Terry E., and Susan C. Duncan. 2009. “The ABC’s of LGM: An Introductory Guide to Latent Variable Growth Curve Modeling.” Social and Personality Psychology Compass 3 (6): 979–91. https://doi.org/10.1111/j.1751-9004.2009.00224.x.
Eid, Michael, Jana Holtmann, Philip Santangelo, and Ulrich Ebner-Priemer. 2017. “On the Definition of Latent-State-Trait Models with Autoregressive Effects.” European Journal of Psychological Assessment.
Enders, Craig K. 2010. Applied Missing Data Analysis. Guilford Press.
Ewing, Sarah W Feldstein, Genevieve F Dash, Wesley K Thompson, Chase Reuter, Vanessa G Diaz, Andrey Anokhin, Linda Chang, et al. 2022. “Measuring Retention Within the Adolescent Brain Cognitive Development (ABCD) SM Study.” Developmental Cognitive Neuroscience 54: 101081.
Fan, Chun Chieh, Andrew Marshall, Harry Smolker, Marybel R Gonzalez, Susan F Tapert, Deanna M Barch, Elizabeth Sowell, et al. 2021. “Adolescent Brain Cognitive Development (ABCD) Study Linked External Data (LED): Protocol and Practices for Geocoding and Assignment of Environmental Data.” Developmental Cognitive Neuroscience 52: 101030.
Feingold, Alan. 2009. “Effect Sizes for Growth-Modeling Analysis for Controlled Clinical Trials in the Same Metric as for Classical Analysis.” Psychological Methods 14 (1): 43.
———. 2019. “Time-Varying Effect Sizes for Quadratic Growth Models in Multilevel and Latent Growth Modeling.” Structural Equation Modeling: A Multidisciplinary Journal 26 (3): 418–29.
French, David P., and Stephen Sutton. 2010. “Reactivity of Measurement in Health Psychology: How Much of a Problem Is It? What Can Be Done about It?” British Journal of Health Psychology 15 (3): 453–68.
Garavan, H, H Bartsch, K Conway, A Decastro, RZ Goldstein, S Heeringa, T Jernigan, A Potter, W Thompson, and D Zahs. 2018. “Recruiting the ABCD Sample: Design Considerations and Procedures.” Developmental Cognitive Neuroscience 32: 16–22.
Geiser, Christian, and Ginger Lockhart. 2012. “A Comparison of Four Approaches to Account for Method Effects in Latent State–Trait Analyses.” Psychological Methods 17 (2): 255.
Ghisletta, Paolo, and John J McArdle. 2012. “Latent Curve Models and Latent Change Score Models Estimated in r.” Structural Equation Modeling: A Multidisciplinary Journal 19 (4): 651–82.
Gonzalez, Raul, Erin L Thompson, Mariana Sanchez, Amanda Morris, Marybel R Gonzalez, Sarah W Feldstein Ewing, Michael J Mason, et al. 2021. “An Update on the Assessment of Culture and Environment in the ABCD Study: Emerging Literature and Protocol Updates over Three Measurement Waves.” Developmental Cognitive Neuroscience 52: 101021.
Graham, John W. 2009. “Missing Data Analysis: Making It Work in the Real World.” Annual Review of Psychology 60 (1): 549–76. https://doi.org/10.1146/annurev.psych.58.110405.085530.
Guillaume, Bryan, Xue Hua, Paul M Thompson, Lourens Waldorp, Thomas E Nichols, Alzheimer’s Disease Neuroimaging Initiative, et al. 2014. “Fast and Accurate Modelling of Longitudinal and Repeated Measures Neuroimaging Data.” Neuroimage 94: 287–302.
Hagler, Donald J., SeanN. Hatton, M. Daniela Cornejo, Carolina Makowski, Damien A. Fair, Anthony Steven Dick, Matthew T. Sutherland, et al. 2019. “Image Processing and Analysis Methods for the Adolescent Brain Cognitive Development Study.” NeuroImage 202 (November): 116091. https://doi.org/10.1016/j.neuroimage.2019.116091.
Hair Jr, Joseph F, G Tomas M Hult, Christian M Ringle, Marko Sarstedt, Nicholas P Danks, Soumya Ray, Joseph F Hair, et al. 2021. “An Introduction to Structural Equation Modeling.” Partial Least Squares Structural Equation Modeling (PLS-SEM) Using R: A Workbook, 1–29.
Hamaker, Ellen L., Rebecca M. Kuiper, and Raoul P. P. P. Grasman. 2015. “A Critique of the Cross-Lagged Panel Model.” Psychological Methods 20 (1): 102–16. https://doi.org/10.1037/a0038889.
Hasan, Alkomiet, Rupert von Keller, Chris Maria Friemel, Wayne Hall, Miriam Schneider, Dagmar Koethe, F Markus Leweke, Wolfgang Strube, and Eva Hoch. 2020. “Cannabis Use and Psychosis: A Review of Reviews.” European Archives of Psychiatry and Clinical Neuroscience 270: 403–12.
Hawes, Samuel W, Amy L Byrd, Raul Gonzalez, Caitlin Cavanagh, Jordan Bechtold, Donald R Lynam, and Dustin A Pardini. 2018. “The Developmental Course of Psychopathic Features: Investigating Stability, Change, and Long-Term Outcomes.” Journal of Research in Personality 77: 83–89.
Hawes, Samuel W, Susan B Perlman, Amy L Byrd, Adrian Raine, Rolf Loeber, and Dustin A Pardini. 2016. “Chronic Anger as a Precursor to Adult Antisocial Personality Features: The Moderating Influence of Cognitive Control.” Journal of Abnormal Psychology 125 (1): 64.
Hedeker, Donald, and Robert D Gibbons. 2006. Longitudinal Data Analysis. Wiley-Interscience.
Hill, Karl G., Danielle Woodward, Tiffany Woelfel, J. David Hawkins, and Sara Green. 2016. “Planning for Long-Term Follow-up: Strategies Learned from Longitudinal Studies.” Prevention Science 17 (7): 806–18.
Hoffman, Lesa. 2015. Longitudinal Analysis: Modeling Within-Person Fluctuation and Change. Routledge.
Hosmer Jr, David W, Stanley Lemeshow, and Susanne May. 2008. Applied Survival Analysis: Regression Modeling of Time-to-Event Data. Vol. 618. John Wiley & Sons.
Howard, Matt C, and Michael E Hoffman. 2018. “Variable-Centered, Person-Centered, and Person-Specific Approaches: Where Theory Meets the Method.” Organizational Research Methods 21 (4): 846–76.
Infante, MA, SC Eberson, Y Zhang, T Brumback, SA Brown, IM Colrain, FC Baker, et al. 2022. “Adolescent Binge Drinking Is Associated with Accelerated Decline of Gray Matter Volume.” Cerebral Cortex 32 (12): 2611–20.
Ji, Zhe, Lin Yuan, Xiong Lu, Hanqing Ding, Jia Luo, and Zun-Ji Ke. 2018. “Binge Alcohol Exposure Causes Neurobehavioral Deficits and GSK3\beta Activation in the Hippocampus of Adolescent Rats.” Scientific Reports 8 (1): 3088.
Kenny, David A., and Alex Zautra. 2001. “Trait–State Models for Longitudinal Data.” In New Methods for the Analysis of Change, edited by Linda M. Collins and A. G. Sayer, 243–63. Washington, DC, US: American Psychological Association. https://doi.org/10.1037/10409-008.
Kievit, Rogier A, Andreas M Brandmaier, Gabriel Ziegler, Anne-Laura van Harmelen, Susanne MM de Mooij, Michael Moutoussis, Ian M Goodyer, et al. 2018. “Developmental Cognitive Neuroscience Using Latent Change Score Models: A Tutorial and Applications.” Developmental Cognitive Neuroscience 33: 99–117.
Kincaid, C. 2005. “Guidelines for Selecting the Covariance Structure in Mixed Model Analysis, Paper 198-30 in Proceedings of the Thirtieth Annual SAS Users Group Conference.” Inc., Cary, North Carolina.
King, Kevin M., Andrew K. Littlefield, Connor J. McCabe, Kathryn L. Mills, John Flournoy, and Laurie Chassin. 2018. “Longitudinal Modeling in Developmental Neuroimaging Research: Common Challenges, and Solutions from Developmental Psychology.” Developmental Cognitive Neuroscience, Methodological Challenges in Developmental Neuroimaging: Contemporary Approaches and Solutions, 33 (October): 54–72. https://doi.org/10.1016/j.dcn.2017.11.009.
Kulisch, Leonard Konstantin, Kadri Arumäe, Daniel A Briley, and Uku Vainik. 2023. “Triangulating Causality Between Childhood Obesity and Neurobehavior: Behavioral Genetic and Longitudinal Evidence.” Developmental Science, e13392.
Lees, Briana, Lindsay M Squeglia, Florence J Breslin, Wesley K Thompson, Susan F Tapert, and Martin P Paulus. 2020. “Screen Media Activity Does Not Displace Other Recreational Activities Among 9–10 Year-Old Youth: A Cross-Sectional ABCD Study.” BMC Public Health 20 (1): 1–11.
Lenz, Sylvia Tamara. 2016. “Alan Agresti (2013): Categorical Data Analysis.” Statistical Papers 57 (3): 849.
Lindquist, Martin A. 2012. “Functional Causal Mediation Analysis with an Application to Brain Connectivity.” Journal of the American Statistical Association 107 (500): 1297–1309.
Lisdahl, Krista M, Susan Tapert, Kenneth J Sher, Raul Gonzalez, Sara Jo Nixon, Sarah W Feldstein Ewing, Kevin P Conway, et al. 2021. “Substance Use Patterns in 9-10 Year Olds: Baseline Findings from the Adolescent Brain Cognitive Development (ABCD) Study.” Drug and Alcohol Dependence 227: 108946.
Little, Roderick J., and Donald B. Rubin. 1989. “The Analysis of Social Science Data with Missing Values.” Sociological Methods & Research 18 (2-3): 292–326. https://doi.org/10.1177/0049124189018002004.
Little, Todd D. 2013. The Oxford Handbook of Quantitative Methods, Vol. 2: Statistical Analysis. Oxford University Press.
Littlefield, Andrew K. 2023. “Longitudinal Data Analysis.” In APA Handbook of Research Methods in Psychology: Data Analysis and Research Publication, Vol. 3, edited by H. E. Cooper, M. N. Coutanche, L. M. McMullen, A. T. Panter, D. E. Rindskopf, and K. J. Sher, 245–68. American Psychological Association.
Littlefield, Andrew K., Kevin M. King, Samuel F. Acuff, Katherine T. Foster, James G. Murphy, and Katie Witkiewitz. 2021. “Limitations of Cross-Lagged Panel Models in Addiction Research and Alternative Models: An Empirical Example Using Project MATCH.” Psychology of Addictive Behaviors. https://doi.org/10.1037/adb0000750.
Littlefield, Andrew K, Kevin M King, Samuel F Acuff, Katherine T Foster, James G Murphy, and Katie Witkiewitz. 2022. “Limitations of Cross-Lagged Panel Models in Addiction Research and Alternative Models: An Empirical Example Using Project MATCH.” Psychology of Addictive Behaviors 36 (3): 271.
Littlefield, Andrew K, Kenneth J Sher, and Douglas Steinley. 2010. “Developmental Trajectories of Impulsivity and Their Association with Alcohol Use and Related Outcomes During Emerging and Young Adulthood i.” Alcoholism: Clinical and Experimental Research 34 (8): 1409–16.
Liu, Yu, Roger E Millsap, Stephen G West, Jenn-Yun Tein, Rika Tanaka, and Kevin J Grimm. 2017. “Testing Measurement Invariance in Longitudinal Data with Ordered-Categorical Measures.” Psychological Methods 22 (3): 486.
Loughnan, Robert J., Clare E. Palmer, Wesley K. Thompson, Anders M. Dale, Terry L. Jernigan, and Chun Chieh Fan. 2020. “Polygenic Score of Intelligence Is More Predictive of Crystallized Than Fluid Performance Among Children.” bioRxiv, 637512.
Lubke, Gitta H, and Bengt Muthén. 2005. “Investigating Population Heterogeneity with Factor Mixture Models.” Psychological Methods 10 (1): 21.
Lubke, Gitta, and Bengt O Muthén. 2007. “Performance of Factor Mixture Models as a Function of Model Size, Covariate Effects, and Class-Specific Parameters.” Structural Equation Modeling: A Multidisciplinary Journal 14 (1): 26–47.
Luciana, M., J. M. Bjork, B. J. Nagel, D. M. Barch, R. Gonzalez, S. J. Nixon, and M. T. Banich. 2018. “Adolescent Neurocognitive Development and Impacts of Substance Use: Overview of the Adolescent Brain Cognitive Development (ABCD) Baseline Neurocognition Battery.” Developmental Cognitive Neuroscience 32: 67–79.
MacKinnon, David P, Amanda J Fairchild, and Matthew S Fritz. 2007. “Mediation Analysis.” Annu. Rev. Psychol. 58: 593–614.
Martz, Meghan E, Mary M Heitzeg, Krista M Lisdahl, Christine C Cloak, Sarah W Feldstein Ewing, Raul Gonzalez, Frank Haist, et al. 2022. “Individual-, Peer-, and Parent-Level Substance Use-Related Factors Among 9-and 10-Year-Olds from the ABCD Study: Prevalence Rates and Sociodemographic Differences.” Drug and Alcohol Dependence Reports 3: 100037.
Masten, Ann S, Glenn I Roisman, Jeffrey D Long, Keith B Burt, Jelena Obradović, Jennifer R Riley, Kristen Boelcke-Stennes, and Auke Tellegen. 2005. “Developmental Cascades: Linking Academic Achievement and Externalizing and Internalizing Symptoms over 20 Years.” Developmental Psychology 41 (5): 733.
Maullin-Sapey, Thomas, and Thomas E Nichols. 2021. “Fisher Scoring for Crossed Factor Linear Mixed Models.” Statistics and Computing 31 (5): 53.
———. 2022. “BLMM: Parallelised Computing for Big Linear Mixed Models.” NeuroImage 264: 119729.
McArdle, JJ, and John R Nesselroade. 1994. “Using Multivariate Data to Structure Developmental Change.” Life-Span Developmental Psychology: Methodological Contributions, 223–67.
McArdle, John J. 2009. “Latent Variable Modeling of Differences and Changes with Longitudinal Data.” Annual Review of Psychology 60: 577–605.
McArdle, John J, and Fumiaki Hamagami. 2001. “Latent Difference Score Structural Models for Linear Dynamic Analyses with Incomplete Longitudinal Data.”
McNeish, Daniel, and Jeffrey R Harring. 2021. “Improving Convergence in Growth Mixture Models Without Covariance Structure Constraints.” Statistical Methods in Medical Research 30 (4): 994–1012.
McNeish, Daniel, and Tyler Matta. 2018. “Differentiating Between Mixed-Effects and Latent-Curve Approaches to Growth Modeling.” Behavior Research Methods 50: 1398–1414.
Mehta, Paras D, and Michael C Neale. 2005. “People Are Variables Too: Multilevel Structural Equations Modeling.” Psychological Methods 10 (3): 259.
Mewton, Louise, Sarah Davies, Matthew Sunderland, Katrina Champion, Nicholas Hoy, Nicola Newton, Maree Teesson, and Lindsay M Squeglia. 2023. “Longitudinal Relationships Between Lifestyle Risk Factors and Neurodevelopment in Early Adolescence.” Health Psychology.
Moffitt, Terrie E. 2015. “Life-Course-Persistent Versus Adolescence-Limited Antisocial Behavior.” Developmental Psychopathology: Volume Three: Risk, Disorder, and Adaptation, 570–98.
Morin, Alexandre JS, Aleksandra Bujacz, and Marylène Gagné. 2018. “Person-Centered Methodologies in the Organizational Sciences: Introduction to the Feature Topic.” Organizational Research Methods. Sage Publications Sage CA: Los Angeles, CA.
Morris, Scott B., and Richard P. DeShon. 2002. “Combining Effect Size Estimates in Meta-Analysis with Repeated Measures and Independent-Groups Designs.” Psychological Methods 7 (1): 105.
Muthén, Bengt, and Linda K Muthén. 2000. “Integrating Person-Centered and Variable-Centered Analyses: Growth Mixture Modeling with Latent Trajectory Classes.” Alcoholism: Clinical and Experimental Research 24 (6): 882–91.
Muthén, L. K. 2017. “Mplus User’s Guide. Los Angeles: Muthén & Muthén; 1998.”
Nweze, Tochukwu, Tobias Banaschewski, Cyracius Ajaelu, Chukwuemeka Okoye, Michael Ezenwa, Robert Whelan, Dimitri Papadopoulos Orfanos, et al. 2023. “Trajectories of Cortical Structures Associated with Stress Across Adolescence: A Bivariate Latent Change Score Approach.” Journal of Child Psychology and Psychiatry.
Orth, Ulrich, D. Angus Clark, M. Brent Donnellan, and Richard W. Robins. 2021. “Testing Prospective Effects in Longitudinal Research: Comparing Seven Competing Cross-Lagged Models.” Journal of Personality and Social Psychology 120 (4): 1013.
Palmer, Clare E, Diliana Pecheva, John R Iversen, Donald J Hagler Jr, Leo Sugrue, Pierre Nedelec, Chun Chieh Fan, Wesley K Thompson, Terry L Jernigan, and Anders M Dale. 2022. “Microstructural Development from 9 to 14 Years: Evidence from the ABCD Study.” Developmental Cognitive Neuroscience 53: 101044.
Parekh, Pravesh, Chun Chieh Fan, Oleksandr Frei, Clare E Palmer, Diana M Smith, Carolina Makowski, John R Iversen, et al. 2021. “FEMA: Fast and Efficient Mixed-Effects Algorithm for Large Sample Whole-Brain Imaging Data.” BioRxiv, 2021–10.
Potter, Alexandra S, Sarahjane L Dube, Lisa C Barrios, Susan Bookheimer, Abigail Espinoza, Sarah W Feldstein Ewing, Edward G Freedman, et al. 2022. “Measurement of Gender and Sexuality in the Adolescent Brain Cognitive Development (ABCD) Study.” Developmental Cognitive Neuroscience 53: 101057.
Preacher, Kristopher J. 2018. “Latent Growth Curve Models.” In The Reviewer’s Guide to Quantitative Methods in the Social Sciences, 178–92. Routledge.
Preacher, Kristopher J, Zhen Zhang, and Michael J Zyphur. 2011. “Alternative Methods for Assessing Mediation in Multilevel Data: The Advantages of Multilevel SEM.” Structural Equation Modeling 18 (2): 161–82.
Preacher, Kristopher J, Michael J Zyphur, and Zhen Zhang. 2010. “A General Multilevel SEM Framework for Assessing Multilevel Mediation.” Psychological Methods 15 (3): 209.
Ramsay, James O, and Bernard W Silverman. 2002. Applied Functional Data Analysis: Methods and Case Studies. Springer.
Rapuano, Kristina M, Nia Berrian, Arielle Baskin-Sommers, Léa Décarie-Spain, Sandeep Sharma, Stephanie Fulton, BJ Casey, and Richard Watts. 2022. “Longitudinal Evidence of a Vicious Cycle Between Nucleus Accumbens Microstructure and Childhood Weight Gain.” Journal of Adolescent Health 70 (6): 961–69.
Raykov, Tenko, and George A Marcoulides. 2012. A First Course in Structural Equation Modeling. routledge.
Ren, Junting, Susan Tapert, Chun Chieh Fan, and Wesley K Thompson. 2022. “A Semi-Parametric Bayesian Model for Semi-Continuous Longitudinal Data.” Statistics in Medicine 41 (13): 2354–74.
Rizopoulos, Dimitris. 2012. Joint Models for Longitudinal and Time-to-Event Data: With Applications in r. CRC press.
Roberts, Brent W, and Wendy F DelVecchio. 2000. “The Rank-Order Consistency of Personality Traits from Childhood to Old Age: A Quantitative Review of Longitudinal Studies.” Psychological Bulletin 126 (1): 3.
Roberts, Brent W, Kate E Walton, and Wolfgang Viechtbauer. 2006. “Patterns of Mean-Level Change in Personality Traits Across the Life Course: A Meta-Analysis of Longitudinal Studies.” Psychological Bulletin 132 (1): 1.
Robins, Lee. 1985. “Epidemiology: Reflections on Testing the Validity of Psychiatric Interviews JAMA Psychiatry JAMA Network.” https://jamanetwork.com/journals/jamapsychiatry/article-abstract/493658.
Rogosa, David, David Brandt, and Michele Zimowski. 1982. “A Growth Curve Approach to the Measurement of Change.” Psychological Bulletin 92 (3): 726.
Rogosch, Fred A, Assaf Oshri, and Dante Cicchetti. 2010. “From Child Maltreatment to Adolescent Cannabis Abuse and Dependence: A Developmental Cascade Model.” Development and Psychopathology 22 (4): 883–97.
Romer, Adrienne L, and Diego A Pizzagalli. 2021. “Is Executive Dysfunction a Risk Marker or Consequence of Psychopathology? A Test of Executive Function as a Prospective Predictor and Outcome of General Psychopathology in the Adolescent Brain Cognitive Development Study.” Developmental Cognitive Neuroscience 51: 100994.
Roy, Ethan, Adam Richie-Halford, John Kruper, Manjari Narayan, David Bloom, Pierre Nedelec, Andreas M Rauschecker, et al. 2024. “White Matter and Literacy: A Dynamic System in Flux.” Developmental Cognitive Neuroscience 65: 101341.
Salthouse, Timothy A. 2014. “Why Are There Different Age Relations in Cross-Sectional and Longitudinal Comparisons of Cognitive Functioning?” Current Directions in Psychological Science 23 (4): 252–56.
Sanchez-Alonso, Sara, and Richard N Aslin. 2020. “Predictive Modeling of Neurobehavioral State and Trait Variation Across Development.” Developmental Cognitive Neuroscience 45: 100855.
Schulenberg, John, Julie Maslowsky, Megan E Patrick, and Meghan Martz. 2019. “Substance Use in the Context of Adolescent Development.” In The Oxford Handbook of Adolescent Substance Abuse, edited by Robert A Zucker and Sandra A Brown, 19–36. Oxford University Press.
Semple, David M, Andrew M McIntosh, and Stephen M Lawrie. 2005. “Cannabis as a Risk Factor for Psychosis: Systematic Review.” Journal of Psychopharmacology 19 (2): 187–94.
Serang, Sarfaraz, Kevin J Grimm, and Zhiyong Zhang. 2019. “On the Correspondence Between the Latent Growth Curve and Latent Change Score Models.” Structural Equation Modeling: A Multidisciplinary Journal 26 (4): 623–35.
Sher, Kenneth J, Kristina M Jackson, and Douglas Steinley. 2011. “Alcohol Use Trajectories and the Ubiquitous Cat’s Cradle: Cause for Concern?” Journal of Abnormal Psychology 120 (2): 322.
Shrout, Patrick E, Gertraud Stadler, Sean P Lane, M Joy McClure, Grace L Jackson, Frederick D Clavél, Masumi Iida, Marci EJ Gleason, Joy H Xu, and Niall Bolger. 2018. “Initial Elevation Bias in Subjective Reports.” Proceedings of the National Academy of Sciences 115 (1): E15–23.
Singer, Judith D. 1998. “Using SAS PROC MIXED to Fit Multilevel Models, Hierarchical Models, and Individual Growth Models.” Journal of Educational and Behavioral Statistics 23 (4): 323–55.
Spear, Linda Patia. 2016. “Consequences of Adolescent Use of Alcohol and Other Drugs: Studies Using Rodent Models.” Neuroscience & Biobehavioral Reviews 70: 228–43.
Stadtbaeumer, Nele, Stefanie Kreissl, and Axel Mayer. 2022. “Comparing Revised Latent State–Trait Models Including Autoregressive Effects.” Psychological Methods.
Steyer, Rolf, Dieter Ferring, and Manfred J Schmitt. 1992. “States and Traits in Psychological Assessment.” European Journal of Psychological Assessment.
Steyer, Rolf, Christian Geiser, and Christiane Loßnitzer. 2023. “Latent State–Trait Models.”
Steyer, Rolf, Axel Mayer, Christian Geiser, and David A Cole. 2015. “A Theory of States and Traits—Revised.” Annual Review of Clinical Psychology 11: 71–98.
Steyer, Rolf, Manfred Schmitt, and Michael Eid. 1999. “Latent State–Trait Theory and Research in Personality and Individual Differences.” European Journal of Personality 13 (5): 389–408.
Sullivan, Edith V., Ty Brumback, Susan F. Tapert, Devin Prouty, Rosemary Fama, Wesley K. Thompson, Sandra A. Brown, Kevin Cummins, Ian M. Colrain, and Fiona C. Baker. 2017. “Effects of Prior Testing Lasting a Full Year in NCANDA Adolescents: Contributions from Age, Sex, Socioeconomic Status, Ethnicity, Site, Family History of Alcohol or Drug Abuse, and Baseline Performance.” Developmental Cognitive Neuroscience 24: 72–83.
Thompson, Wesley K., Deanna M. Barch, James M. Bjork, Raul Gonzalez, Bonnie J. Nagel, Sara Jo Nixon, and Monica Luciana. 2019. “The Structure of Cognition in 9 and 10 Year-Old Children and Associations with Problem Behaviors: Findings from the ABCD Study’s Baseline Neurocognitive Battery.” Developmental Cognitive Neuroscience 36: 100606.
Thompson, Wesley K, Joachim Hallmayer, Ruth O’Hara, and Alzheimer’s Disease Neuroimaging Initiative. 2011. “Design Considerations for Characterizing Psychiatric Trajectories Across the Lifespan: Application to Effects of APOE-\varepsilon4 on Cerebral Cortical Thickness in Alzheimer’s Disease.” American Journal of Psychiatry 168 (9): 894–903.
Thompson, Wesley K, Gauri N Savla, Ipsit V Vahia, Colin A Depp, Ruth O’Hara, Dilip V Jeste, and Barton W Palmer. 2013. “Characterizing Trajectories of Cognitive Functioning in Older Adults with Schizophrenia: Does Method Matter?” Schizophrenia Research 143 (1): 90–96.
Trevino, Angel D, Belal Jamil, Jinni Su, Fazil Aliev, Kit K Elam, and Kathryn Lemery-Chalfant. 2023. “Alcohol Use Disorder Polygenic Risk Scores and Trajectories of Early Adolescent Externalizing Behaviors: Examining the Role of Parenting and Family Conflict in the Racially/Ethnically Diverse ABCD Sample.” Behavior Genetics, 1–18.
Twisk, Jos, Michiel de Boer, Wieke de Vente, and Martijn Heymans. 2013. “Multiple Imputation of Missing Values Was Not Necessary Before Performing a Longitudinal Mixed-Model Analysis.” Journal of Clinical Epidemiology 66 (9): 1022–28.
Uban, Kristina A., Megan K. Horton, Joanna Jacobus, Charles Heyser, Wesley K. Thompson, Susan F. Tapert, Pamela A. F. Madden, and Elizabeth R. Sowell. 2018. “Biospecimens and the ABCD Study: Rationale, Methods of Collection, Measurement and Early Data.” Developmental Cognitive Neuroscience, The Adolescent Brain Cognitive Development (ABCD) Consortium: Rationale, Aims, and Assessment Strategy, 32 (August): 97–106. https://doi.org/10.1016/j.dcn.2018.03.005.
Van De Schoot, Rens, Peter Schmidt, Alain De Beuckelaer, Kimberley Lek, and Marielle Zondervan-Zwijnenburg. 2015. “Measurement Invariance.” Frontiers in Psychology. Frontiers Media SA.
Van Dijk, Milenna T, Eleanor Murphy, Jonathan E Posner, Ardesheer Talati, and Myrna M Weissman. 2021. “Association of Multigenerational Family History of Depression with Lifetime Depressive and Other Psychiatric Disorders in Children: Results from the Adolescent Brain Cognitive Development (ABCD) Study.” JAMA Psychiatry 78 (7): 778–87.
VanderWeele, Tyler J. 2016. “Mediation Analysis: A Practitioner’s Guide.” Annual Review of Public Health 37: 17–32.
VanderWeele, Tyler J, and Eric J Tchetgen Tchetgen. 2017. “Mediation Analysis with Time Varying Exposures and Mediators.” Journal of the Royal Statistical Society Series B: Statistical Methodology 79 (3): 917–38.
Vize, Colin, Whitney R Ringwald, Emily R Perkins, Rebecca Waller, Samuel Hawes, and Amy L Byrd. 2023. “The Hierarchical Structure and Longitudinal Measurement Invariance of Externalizing Symptoms in the Adolescent Brain and Cognitive Development (ABCD) Study.”
Volkow, Nora D., George F. Koob, Robert T. Croyle, Diana W. Bianchi, Joshua A. Gordon, Walter J. Koroshetz, Eliseo J. Pérez-Stable, et al. 2018. “The Conception of the ABCD Study: From Substance Use to a Broad NIH Collaboration.” Developmental Cognitive Neuroscience, The Adolescent Brain Cognitive Development (ABCD) Consortium: Rationale, Aims, and Assessment Strategy, 32 (August): 4–7. https://doi.org/10.1016/j.dcn.2017.10.002.
Wang, Lijuan, Qian Zhang, Scott E Maxwell, and CS Bergeman. 2019. “On Standardizing Within-Person Effects: Potential Problems of Global Standardization.” Multivariate Behavioral Research 54 (3): 382–403.
Watson, Nicole, Eva Leissou, Heidi Guyer, and Mark Wooden. 2018. “Best Practices for Panel Maintenance and Retention.” In Advances in Comparative Survey Methods, 597–622. John Wiley & Sons, Ltd. https://doi.org/10.1002/9781118884997.ch29.
Wiker, Thea, Mads L Pedersen, Lia Ferschmann, Dani Beck, Linn B Norbom, Andreas Dahl, Tilmann von Soest, et al. 2023. “Assessing the Longitudinal Associations Between Decision-Making Processes and Attention Problems in Early Adolescence.” Research on Child and Adolescent Psychopathology, 1–15.
Willoughby, Michael T, RJ Wirth, and Clancy B Blair. 2012. “Executive Function in Early Childhood: Longitudinal Measurement Invariance and Developmental Change.” Psychological Assessment 24 (2): 418.
Woo, Sang Eun, Joeri Hofmans, Bart Wille, and Louis Tay. 2024. “Person-Centered Modeling: Techniques for Studying Associations Between People Rather Than Variables.” Annual Review of Organizational Psychology and Organizational Behavior 11: 453–80.
Wood, Simon N. 2017. Generalized Additive Models: An Introduction with r. chapman; hall/CRC.
Xiang, Qu, Kai Chen, Li Peng, Jiawei Luo, Jingwen Jiang, Yang Chen, Lan Lan, Huan Song, and Xiaobo Zhou. 2022. “Prediction of the Trajectories of Depressive Symptoms Among Children in the Adolescent Brain Cognitive Development (ABCD) Study Using Machine Learning Approach.” Journal of Affective Disorders 310: 162–71.
Zhao, Yi, Xi Luo, Martin Lindquist, and Brian Caffo. 2018. “Functional Mediation Analysis with an Application to Functional Magnetic Resonance Imaging Data.” arXiv Preprint arXiv:1805.06923.
Zucker, Robert A., Raul Gonzalez, Sarah W. Feldstein Ewing, Martin P. Paulus, Judith Arroyo, Andrew Fuligni, Amanda Sheffield Morris, Mariana Sanchez, and Thomas Wills. 2018. “Assessment of Culture and Environment in the Adolescent Brain and Cognitive Development Study: Rationale, Description of Measures, and Early Data.” Developmental Cognitive Neuroscience 32: 107–20.