Introduction

Adolescence is a time of increased risky behaviors, including initiation and escalation of alcohol and other drug use [13]. Early substance use initiation is strongly associated with the development of a substance use disorder (SUD), with lifetime rates of alcohol dependence estimated at over 40 % [4] and drug dependence at over 30 % [5] among those who begin use by age 13. Parental SUD further raises risk for offspring SUD beyond risk conferred by age of first use [4, 5]. SUD is a serious health concern, having been linked to a wide range of mental, physical, and social harms [68]. In order to develop more effective prevention and intervention strategies to reduce the morbidity and mortality associated with SUD, understanding the mechanisms that precede the onset and escalation of use is crucial.

Over the past decade, there has been an emergence of studies investigating the neural basis of risk for SUD using functional magnetic resonance imaging (fMRI), much of which has focused on the biological underpinnings of risk-taking behaviors that facilitate substance use. In general, these include neural mechanisms underlying inhibitory control, which involves executive functioning and can be considered a top-down system supporting behavioral control, and incentive reactivity, which responds to rewarding or salient stimuli and can be considered a more reactive, or bottom-up system [911]. This review summarizes recent findings (i.e., within the last 3 years) regarding neurofunctional risk markers for SUD in children, adolescents, and emerging adults, specifically in the domains of inhibitory control and reward responsiveness. There are many ways to approach the investigation of risk; we focus this review on family history, prospective, and imaging genetic studies. Gender effects, when analyzed, are noted. We discuss convergences in the literature and highlight findings suggesting that relationships between SUD risk and neurofunctioning may be moderated by age, gender, and history of substance use. Recommendations for future directions are also discussed.

Inhibitory Control

Poor inhibitory control is a key mechanism underlying behavioral undercontrol generally and vulnerability to disinhibitory psychopathology, such as substance abuse, more specifically [12]. In adolescents, impulsivity has been associated with earlier and greater substance use [13, 14] and SUD diagnosis [15]. Given the robust association between inhibitory control and risk for SUD, there has been great interest in uncovering the neural substrates that may mediate this relationship. The following sections review family history and prospective studies examining brain regions and circuits associated with response inhibition and substance use behavior.

Family History Studies

Two recent studies have examined the influence of family history on inhibitory circuitry. Silveri et al. [16] investigated the impact of parental alcohol use disorder (AUD) on brain activation during a Stroop task in healthy 8–19-year olds with fewer than four lifetime episodes of alcohol or drug use. During Stroop interference, youth with parental AUD (FH+) showed greater activation of the frontal network involved in response inhibition, including middle frontal and cingulate gyri, than youth without parental AUD (FH−). This was interpreted as reflecting compensatory mechanisms due to reduced neuronal efficiency of this network. In contrast, findings from earlier cross-sectional work in children and adolescents suggest that blunted frontal activation during inhibitory control may be an underlying factor in risk [17, 18] (although this was not apparent in a sample of late adolescents/emerging adults [19]). One reason for this difference may be that the Stroop task is more cognitively demanding than other inhibitory tasks, thereby unmasking functional differences under more effortful conditions. Furthermore, the ability to successfully inhibit a response increases from childhood into early adulthood [20], concomitant with maturation of inhibitory control circuitry [21]. Therefore, the use of a broad age range (8–19 years) may mask important developmental differences.

In order to address developmental differences in inhibitory control functioning among at-risk youth, Hardee et al. [22] used a longitudinal study design. Youth entered the study at 7–12 years of age and underwent 1–3 follow-up scans at 1- to 2-year intervals, covering the age range of 7–17 years. Inhibitory control was probed with a go/no-go task, which engages individuals in responding to frequent “go” signals and requires them to inhibit a prepotent response when infrequent “no-go” signals occur. Significant age-by-group interactions were found in right caudate, middle cingulate, and middle frontal gyrus during successful no-go trials. At baseline, when participants were 7–12 years old and had minimal substance use, FH+ youth had blunted activation in each of these regions compared with FH− youth, consistent with prior evidence [17, 18]. With age, activation decreased in the caudate and middle frontal gyrus in FH− youth only, suggesting a normal developmental pattern absent in high-risk youth. In contrast, FH+ youth demonstrated increased middle cingulate activation with age, interpreted as a compensatory mechanism for a weakness or developmental delay in response inhibition circuitry; thus, more effort may be required in high-risk youth to overcome the prepotent automatic response generated in the go/no-go task.

Prospective Studies

A recent emergence of prospective neuroimaging studies has identified early-occurring brain functional differences during inhibitory control tasks that differentiate youth who do and do not go on to problem substance use. Using baseline fMRI data as a predictor of later substance use or problems, prospective studies provide a more definitive connection between neural activation patterns and risk than possible with family history studies.

Blunted Activation During Inhibitory Control as a Risk Marker for Problem Substance Use

Norman et al. [23•] scanned 12–14-year olds with minimal substance use during a go/no-go task and categorized them as heavy drinkers or controls approximately 4 years later. Youth who transitioned into heavy alcohol use displayed blunted neural activation during no-go trials at baseline, prior to the onset of use, in widespread regions of inhibitory control circuitry, including the frontal, parietal, and temporal cortices and striatum. More recently, Mahmood et al. [24] scanned high- and low-frequency substance users between the ages of 16–19 years during a go/no-go task to investigate whether activation during no-go trials predicted use and dependence symptoms 18 months later. Less ventromedial prefrontal activation predicted more drug and alcohol dependence symptoms at follow-up, above and beyond baseline symptoms and lifetime drug use, particularly in high-frequency users. In contrast, increased engagement of the left angular and supramarginal gyri during no-go trials predicted more total drug use occasions in the follow-up period, again, an effect most evident in high-frequency users. The left angular and supramarginal gyri are not commonly associated with inhibitory control [25, 26]; thus, the authors suggest that, in order to achieve comparable performance, these youth may use an alternate strategy to compensate for a relative weakness in cognitive control. It is unclear from this work whether the emergence of heightened risk-related activation is a developmental effect related to the older age of participants (e.g., compared to [18, 23•]) as the work of Hardee et al. might suggest [22], or whether this activation is due to the inclusion of high-frequency substance users.

Recent longitudinal work by Wetherill et al. [27•] suggests that the initiation of heavy substance use during adolescence may result in less efficient response inhibition circuitry, marked by heightened activation. Baseline fMRI data was acquired during the go/no-go task in 11–16-year olds with minimal substance use and again approximately 3 years later after grouping them into heavy drinkers and demographically matched non-drinking controls. At baseline, youth who later became heavy drinkers showed blunted activation in the bilateral middle frontal gyrus, right inferior parietal lobule, left putamen, and left cerebellum during no-go trials, consistent with prior work [18, 22, 23•, 24]. At follow-up, however, these same youth showed greater activation in each region (excluding the putamen). Therefore, blunted inhibitory control functioning may be associated with risk prior to heavy substance use, whereas after heavy use is initiated, inefficient neural recruitment may begin to play a larger role.

Brain Functioning During Inhibitory Errors as a Risk Marker for Problem Substance Use

Error processing is critical for adaptive adjustment of performance and thus is an important aspect of inhibitory control [28]. Much of the work using the go/no-go task to investigate familial risk either focuses on successful trials only [19, 22] or does not differentiate between successful and failed no-go trials [18, 23•, 24, 27•]. Recent work by Heitzeg et al. [29] extends this literature by differentiating between successful and failed inhibitory (no-go) trials. Participants underwent a baseline scan at age 9–12 years, and those who demonstrated problem substance use by age 13–16 were matched on age, gender, and parental AUD to substance-naïve controls. No baseline activation differences were observed between groups during successful no-go trials. However, those who went on to have problem substance use demonstrated blunted activation in the left middle frontal gyrus to inhibitory errors (failed no-go trials) compared with controls at baseline. Blunted activation to errors in this region at age 9–12 was associated with more externalizing problems at age 11–13 and predicted problem substance use by age 13–16 over and above externalizing problems. Although this sample consisted primarily of males, exploratory analyses revealed that all effects trended in the same direction among females. Thus, blunted activation in error monitoring circuitry at an early age may underlie problems adapting behavior appropriately, leading to undercontrolled behavior and early problem substance use.

Work from IMAGEN, a large-scale longitudinal study of adolescent development [30], has also identified brain functioning during failed inhibition as one important risk factor for problem substance use [31••]. Using machine learning, models of risk were generated incorporating a broad range of factors including personality, life experience, cognition, genetics, brain structure, and brain function. Magnitude of prefrontal activation to inhibitory errors during a stop signal task at age 14 was one significant predictor of binge drinking by age 16.

Summary

The studies reviewed above are summarized in Table 1. Taken together, evidence indicates that in childhood and early adolescence, a blunting of response inhibition circuitry is a risk factor for later problem substance use, whereas more complex patterns emerge later in adolescence as inhibitory control circuitry matures and heavy substance use begins. The evidence converges on risk-related effects in the prefrontal cortex with findings specifically in the middle frontal gyrus common across all studies (with one exception [24]). Results are less consistent regarding other inhibitory control regions. Methodical issues likely contribute to this variability; most studies (with two exceptions [16, 31••]) used a variation of the go/no-go task, yet timing and designs differed across studies and the specific data used in analyses varied based on whether fixation, go trials, or implicit baselines were used and whether successful and failed no-go trials were differentiated.

Table 1 Summary of primary findings in the domains of inhibitory control and reward sensitivity from family history, imaging genetic and prospective studies of risk for substance use disorder in children, adolescents and emerging adults

An emerging literature describes an impact of genetic variation in dopamine signaling on prefrontal cortical function during inhibitory control in typically developing adolescents [32] and young adults [33], and on impulsivity [34] and drug craving [35] in adults. However, thus far, there have been no studies investigating specific relationships among genetic variation, neural functioning during inhibitory control, and risk for problem substance use in children, adolescents, or emerging adults. This will be an important direction for future research.

Gender differences were investigated directly in only one study [29], as an exploratory analysis with a small sample of females. No suggestion of a gender difference in the relationship between brain mechanisms involved in inhibitory control and early risk for problem substance use was found in this 9–12-year-old sample; however, evidence in adults indicates that increased thalamic activation during inhibitory errors uniquely marks parental risk in males but not females [36]. Further work is required investigating gender differences in children and adolescents prior to the onset of significant substance use to determine whether early-occurring neural functioning of inhibitory control systems may relate to risk for SUD differently in males and females.

Reward Sensitivity

Reward sensitivity is especially pronounced during adolescence, aligning with developmental increases in sensation-seeking and substance use that occur during this period [2, 37]. The peak in reward sensitivity during adolescence is associated with reorganization of the dopaminergic system, including projections from the limbic to the prefrontal system [37, 38]. The mesolimbic dopamine system is critical to incentive processing [39], is centrally involved in the reinforcing properties of drugs of abuse [40], and has been implicated in vulnerability to addiction [4143]. Thus, reward system functioning has emerged as an important target in the study of neurofunctional risk markers for SUD. There are competing views, however, regarding the relationship between reward functioning and risk; some hypotheses propose a hyper-responsive system biases individuals toward heightened reward-seeking leading to increased substance use (e.g., [44, 45]), whereas others propose a hypo-responsive system motivates individuals to seek out strong rewards such as drugs to raise the system to normal levels [46]. The following sections review family history, imaging genetic, and prospective studies of the association between reward functioning and substance use behavior.

Family History Studies

Regional Brain Activation in Response to Reward Cues

Several recent studies have investigated the influence of family history on reward system function with mixed evidence for reward-related SUD-risk markers. Stice and Yokum [47] used a coin flip task in 14–17-year olds to study how parental SUD influences the neural circuitry involved in monetary reward anticipation. Participants were matched on gender and had not used substances in the previous year. FH+ youth displayed greater activation in the left dorsolateral prefrontal cortex and bilateral putamen and less activation in the fusiform gyrus and inferior temporal gyrus relative to FH− youth, suggesting that at-risk youth exhibit increased responsiveness but also greater attentional disengagement during reward anticipation. These results support a reward-surfeit model of addiction, whereby those at risk have increased reward system sensitivity. A limitation, however, is the lack of lifetime substance use characterization in the groups. Even in a non-addicted sample, prior drug exposure may modulate reward circuitry response [48, 49], and FH+ youth would be expected to have more lifetime substance use [50, 51].

Yau et al. [52] investigated how parental AUD impacts reward system function in 18–22-year olds who were individually matched across family history groups on lifetime and current substance use, as well as age and gender. The focus was specifically on nucleus accumbens (NAcc) activation during a monetary incentive delay task (MIDT) [53]. The MIDT is a well-validated task used to study neural circuitry involved in incentive reactivity; it has been associated with robust ventral striatal (VS) activation, including the NAcc [54]. Overall, the FH+ group demonstrated blunted NAcc activation during reward anticipation compared with FH− youth. However, uniquely within the FH+ group, NAcc activation positively correlated with earlier externalizing behaviors and current and lifetime amounts of drinking, which belies a simple interpretation relating hypo-activation to SUD risk. Instead, links among heightened NAcc responsivity, substance use, and behavioral risk may underlie vulnerability uniquely in FH+ youth [52]. When investigating heavy versus light drinkers within each group, an interaction between family history and alcohol use was revealed. Specifically, light-drinking FH+ youth had blunted activation compared with their heavy-drinking FH+ counterparts as well as compared with all FH− youth. In contrast, heavy-drinking FH+ youth did not differ from FH− youth. The authors suggest that this blunted NAcc response in light-drinking FH+ youth reflects a resilience mechanism.

Interestingly, in a younger sample, Muller et al. [55] found no differences in activation to reward anticipation or receipt during the MIDT in AUD FH+ versus FH− youth. They investigated a large sample of 13–15-year olds, matched on gender and with similar levels of substance use across groups. This finding is consistent with prior work reporting no differences in NAcc activation during the MIDT in 12–16-year-old AUD FH+ and FH− groups [56]. As the authors noted, it is possible that conferment of familial risk related to reward responsivity emerges later in development [55]. Indeed, some evidence exists for changes in risk-related activation of NAcc to reward across development [57•], described below. However, it is likely that other methodological differences also play a role in discrepant results across studies, such as task design and timing, and subject-selection criteria, including levels of prior substance use and psychopathology.

Functional Imbalance of Reward and Control Regions

Of major significance to adolescent substance use is evidence suggesting that the relatively early maturation of subcortical incentive-reactive systems compared with later-maturing prefrontal control systems may bias adolescents toward seeking out reward [5861]. This bias is proposed to underlie the impulsive, risky decision-making associated with adolescence [62, 63] and may also contribute to heightened risk for SUD; this has been the focus of several recent studies. Ivanov et al. [64] studied drug-naïve 8–13-year olds, primarily male, all with attention deficit/hyperactivity disorder (ADHD) and half with at least one parent with an SUD. An anticipation-conflict-reward task was used, which includes both reward and executive control components. FH+ youth had greater activity in motivational-reward circuitry, including left anterior insula (extending into caudate), left inferior frontal gyrus, and left orbitofrontal cortex, in response to reward anticipation and notification compared with FH− youth. FH+ youth also showed less activity in behavioral inhibition circuitry, including right anterior cingulate cortex and left dorsomedial prefrontal cortex, during cognitive conflict. This study supports the hypothesis that SUD risk is related to a functional imbalance between reward and control areas.

Further support for a model of functional imbalance comes from work by Cservenka and Nagel [65], who investigated brain functioning during reward-related decision-making in 13–15-year-old AUD FH+ and FH− youth with minimal substance use. Risk taking in the context of high- and low-probability reward was examined using a Wheel of Fortune task [66]. Despite comparable risk-taking behavior between groups, activation in dorsolateral prefrontal cortex and cerebellum was blunted during risky choices in FH+ compared with FH− youth, suggesting reduced cognitive control during risky decision-making in at-risk youth.

Connectivity Studies of the Balance Between Reward and Control Circuitry

Two recent connectivity studies also lend support to risk-related differences in the balance between reward and control circuitry. Using resting-state fMRI, Cservenka et al. [67] explored the intrinsic connectivity of the NAcc in 10–16-year-old AUD FH+ and FH− youth with minimal substance use. FH+ youth had reduced segregation between the left NAcc and bilateral inferior frontal gyri compared with FH− youth; reduced segregation was associated with increased sensation-seeking at a trend level. FH+ youth also had negative functional connectivity between the right NAcc and left orbitofrontal cortex, whereas FH− subjects had positive functional connectivity. This suggests that, in at-risk youth, cognitive control and reward networks are not appropriately dissociated, and appetitive brain regions are not appropriately integrated [67]. This is one of the few studies that investigated a possible moderating influence of gender; however, no significant effect was observed.

Weiland et al. [68] probed the impact of parental AUD on task-related functional connectivity of the NAcc during the MIDT in 18–22-year olds. Substance use history did not differ significantly between risk groups. During incentive anticipation, NAcc connectivity with attention, motor, and default mode network regions—including paracentral lobule/precuneus and sensorimotor areas—was decreased in FH− but increased in FH+ youth. In FH+ youth only, connectivity with NAcc in these regions correlated positively with sensation-seeking as well as drinking volume; the strength of connectivity significantly mediated the relationship between sensation-seeking and drinking volume. The authors suggest that differences in NAcc coupling with attention and motor networks may increase risk by influencing motor output in response to incentive cues. Though generally in line with the aforementioned studies, no differences were observed between groups in NAcc coupling with prefrontal cognitive control regions specifically. Instead, this finding is suggestive of more widespread network connectivity dysfunction in high-risk youth than what is predicted by a circumscribed model of imbalance between reward and cognitive control areas.

Imaging Genetic Studies

An emerging literature describes an impact of genetic variation in dopamine signaling on reward-related brain function in healthy adults [6972]. There has been recent interest in investigating how genetic influences on reward system functioning may contribute to risk for problem substance use. The following paragraphs review associations of GAL and GABRA2 in relation to fMRI reward response and substance use behavior in children and emerging adults.

GAL, Reward Processing, and Risk for Problem Substance Use

Galanin is a neuropeptide that has been linked to appetitive and consummatory behaviors, including drinking alcohol, via dopamine stimulation in the VS [73, 74]. Nikolova et al. [75••] tested the association between VS reactivity and a common functional haplotype (GAL5.1) of a gene coding for galanin (GAL) in 138 college students (mean age 19.8 years) using a number-guessing paradigm. Positive feedback during this task elicits significant reward-related VS activity [72]. The GG haplotype, associated with higher GAL promoter activity, was hypothesized to be associated with greater VS reactivity and greater problem drinking than the CA haplotype; however, there was no direct effect of GAL5.1 haplotype on VS reactivity or problem drinking. A genotype-by-gender interaction did emerge, however, with male CA carriers having higher VS activation than female CA carries, and female GG homozygotes higher than female CA carriers. Furthermore, reduced VS reactivity mediated an association between CA haplotype and lesser problem drinking in women, whereas in men, increased VS reactivity mediated the association between the CA haplotype and greater problem drinking. This novel work supports a role of galanin signaling within reward pathways in risk for problem substance use and also highlights an important gender difference in how this risk may unfold.

GABRA2, Reward Processing, and Risk for Problem Substance Use

Two recent studies focused on the impact of GABRA2 on reward processing in relationship to SUD risk [57•, 76]. The GABRA2 gene encodes the α2 subunit of the γ-aminobutyric acid A receptor (GABAA). This subunit is among the predominant receptor subtypes expressed in the NAcc [77] and contributes to inhibitory regulation of NAcc dopaminergic function [78]. Associations have been found between GABRA2 and adult AUD (e.g., [79, 80]) as well as phenotypic precursors to SUD [81]. Villafuerte et al. [76] demonstrated an association between GABRA2 and brain function during the MIDT in 18–22-year olds; those homozygous for the risk allele (G) showed heightened left insula activation during reward anticipation and were more likely to have alcohol dependence symptoms. Post hoc analysis revealed the effect was significant only in females. Both GG genotype and insula activation during reward anticipation were linked to impulsive behavior across the entire sample.

More recently, Heitzeg et al. [57•] examined the influence of GABRA2 on the developmental trajectory of NAcc activation during the MIDT longitudinally. Participants were 76 children (ages 8–13) and 99 emerging adults (ages 18–23) at study entry; 79 % were AUD FH+, characteristic of a high-risk sample. The majority underwent repeated scanning at 1- to 2-year intervals, covering the age range of 8–27. Substance use did not significantly differ by genotype but was used as a covariate to account for greater use with age. A significant age-by-genotype effect was observed whereby risk (G) allele carriers (GG and AG) had heightened NAcc activation to reward anticipation during adolescence but not childhood or young adulthood. In contrast, A homozygotes showed no significant age-related changes. During adolescence, G homozygotes showed significantly greater NAcc activation than A homozygotes, an effect that was not apparent in childhood or young adulthood. Supplementary analyses did not reveal any significant effects of gender. These findings support the view that reward-related SUD risk may be developmentally modulated (e.g., [55]) and represent an important step toward understanding genetic and neural influences underlying the unfolding of risk for SUD across development.

Prospective Studies

To date, there are no reports investigating whether early-occurring brain functional differences during reward, prior to significant substance use, can differentiate youth who do and do not go on to problem substance use. However, three prospective studies have examined relationships between reward circuitry functioning and later problem substance use in already-using samples.

In the same study described previously, Heitzeg et al. [57•] found that increased NAcc activation to reward anticipation during the MIDT was positively associated with the number of alcohol-related problems reported over the next 3–6 years, even when controlling for lifetime drinking at time of the scan. Furthermore, mediation analysis revealed there was a significant indirect effect of GABRA2 genotype on later alcohol problems that was transmitted through NAcc activation.

A prospective study by Dager et al. [82] examined the relationship between the neural response to alcohol pictures and escalation of drinking over the following year in 18–21-year-old college students. Participants were categorized into three groups based on alcohol use patterns at baseline and follow-up assessments: continuously moderate drinkers, continuously heavy drinkers, and transitioners from moderate to heavy drinking. Transitioners had a heightened response to alcohol versus non-alcohol pictures within cue-reactive circuitry (e.g., [83]), including the caudate, ventromedial prefrontal cortex, anterior cingulate, orbitofrontal cortex, and insula. Greater reactivity of this circuitry predicted increased drinking and more alcohol-related problems, beyond baseline level of drinking and self-reported impulsivity. Exploratory analyses found no effects of gender. This supports a model whereby enhanced motivation for alcohol or other drugs over-engages networks involved in reward and motivation [84]. Together, both the Heitzeg et al. [57•] and Dager et al. [82] studies suggest that heightened activity in reward networks early in an individual’s substance-using “career” may mark those at the greatest risk for future problem use.

A third prospective study reported by Whelan et al. [31••] found that heightened activation in superior frontal gyrus to reward outcome during the MIDT at age 14 was one of several significant predictors of binge drinking by age 16 in the IMAGEN sample. These findings lend further support to the view that heightened activity in reward networks can signal future risk. However, these authors also reported a contribution of reduced activation in posterior regions not typically associated with reward responding to risk for later binge drinking.

Summary

The studies reviewed above are summarized in Table 1. Taken together, the bulk of evidence supports a relationship between hyperactive neural responses in reward-related circuitry and risk for SUD. This was supported by family history studies [47, 52, 64], imaging genetic studies [57•, 75••, 76], and prospective studies [31••, 57•, 82], and some evidence indicates the effect may be developmentally modulated [57•]. There is also some suggestion that reduced activation to reward in posterior regions, including occipital and temporal cortices and posterior cingulate, may be a risk factor for SUD [31••, 47], possibly reflecting abnormal processing of, and/or attention to, rewarding stimuli. Furthermore, accumulating evidence suggests that a functional imbalance between reward and control areas may contribute to risk-related hyperactivity of reward system functioning [64, 67, 68], which converges with findings of weakness in inhibitory control mechanisms in at-risk youth. There is some indication of gender differences in these brain functional pathways to risk (e.g., [75••, 76]), although others found no effects [57•, 67, 82], and most studies did not explore gender differences.

Conclusions and Directions for Future Research

The development of SUD is a multistage process, beginning with alcohol or drug experimentation, followed by more frequent use, and finally for a subset of users, compulsive use. Peak 12-month prevalence of SUD occurs between ages 18 and 23 [85, 86]. Therefore, for many, as this cascade of risk unfolds, significant developmental changes are also occurring in personality, behavior, and socialization, and in brain structure and function. Individual differences in cognitive control and reward responsive circuitry may relate to the development and maintenance of SUD differently at various stages in the process (e.g., [87]) and at various developmental periods (e.g., [22, 57•]). Furthermore, the use of substances at a critical period in development may have implications for the functioning of these systems (e.g., [27•]; also see reviews in [8891]) and the likelihood of transitioning into problem use. The perennial challenge, then, is disentangling relationships among risk, substance use, and neurodevelopment.

In recent years, there has been an emergence of longitudinal studies, which allow for powerful prospective approaches to identifying age-specific risk factors and enable characterization of neural systems and emerging substance use and problems in parallel. Additionally, large-scale, multi-site longitudinal studies are allowing for multivariate analyses across a vast range of risk factors to enable broader inferences (e.g., [30, 31••]). Much of this work begins with baseline assessment in early adolescence, which has the likely advantage of capturing emerging substance use within a 5-year period, due to normative increases in use throughout adolescence. A limitation, however, is that youth at the highest risk for SUD have already initiated substance use by early adolescence. Research indicates that substance use by age 13 substantially increases risk for later dependence [4, 5]; thus, longitudinal studies targeting high-risk populations some years prior to this age and following them through adolescence into early adulthood are critical. Furthermore, much of the work reviewed here did not explore gender differences. While rates of past-year SUD in 12–17-year olds are equal between males and females (6.1 %), a significant gender divergence emerges afterward, with higher rates in males aged 18 or older (12.2 %), compared with females (5.7 %) [86]. An understanding of how gender differences in brain development and functioning may contribute to differential risk for SUD will be an important future focus of research.

Findings of moderating effects of family history (e.g., [52, 68]) and gender (e.g., [75••]) illustrate heterogeneity in the relationship between brain function and risk for SUD and highlight the importance of investigating potential moderating and mediating influences that may impact outcome. These include not only a high-risk genetic background marked by family history [92, 93] but also poor parenting [94], stress [95], conflict in the home [96], and peer involvement [97]. For example, recent work has illustrated a moderating effect of the presence of peers on adolescent risk taking, and mediation of this effect by functioning of reward-related circuitry, as well as developmental impacts on the effect [98, 99]. This has not yet been investigated in the specific context of SUD risk. A thorough developmental understanding of vulnerability pathways will require longitudinal characterization of influences across multiple levels, including environmental and social contexts, gender, behavior, neural functioning, and genetics. Knowledge of how these complex systems interact to provide developmental periods of increased or decreased vulnerability has the potential to guide the type and timing of interventions for at-risk individuals to improve outcomes.