Hostname: page-component-cd9895bd7-p9bg8 Total loading time: 0 Render date: 2024-12-26T06:23:39.851Z Has data issue: false hasContentIssue false

Navigating Potential Pitfalls in Difference-in-Differences Designs: Reconciling Conflicting Findings on Mass Shootings’ Effect on Electoral Outcomes

Published online by Cambridge University Press:  12 April 2024

HANS J. G. HASSELL*
Affiliation:
Florida State University, United States
JOHN B. HOLBEIN*
Affiliation:
University of Virginia, United States
*
Hans J. G. Hassell, Professor, Department of Political Science, Florida State University, United States, hans.hassell@fsu.edu.
Corresponding author: John B. Holbein, Associate Professor of Public Policy, Politics, and Education, Frank Batten School of Leadership and Public Policy, University of Virginia, United States, holbein@virginia.edu.
Rights & Permissions [Opens in a new window]

Abstract

Work on the electoral effects of gun violence in the U.S. relying on difference-in-differences designs has produced findings ranging from null to substantively large effects. However, as difference-in-difference designs, on which this research relies, have exploded in popularity, scholars have documented several methodological issues including potential violations of parallel-trends and unaccounted for treatment effect heterogeneity. These pitfalls (and their solutions) have not been fully explored in political science. We apply these advancements to the unresolved debate on gun violence’s effects on U.S. electoral outcomes. We show that studies finding a large positive effect of gun violence on Democratic vote shares are a product of a failure to properly specify difference-in-differences models when underlying assumptions are unlikely to hold. Once these biases are corrected, shootings show little evidence of sparking large electoral change. Our work clarifies an unresolved debate and provides a cautionary guide for scholars currently employing difference-in-differences designs.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press on behalf of American Political Science Association

Gun violence in the United States has a devastating impact (e.g., Barney and Schaffner Reference Barney and Schaffner2019; Hartman and Newman Reference Hartman and Newman2019; Marsh Reference Marsh2022; Rogowski and Tucker Reference Rogowski and Tucker2019; Rossin-Slater et al. Reference Rossin-Slater, Schnell, Schwandt, Trejo and Uniat2020). Yet, despite repeated tragedies and public support for policies to reduce gun violence, policy response has been tepid (Goss Reference Goss2010; Luca, Malhotra, and Poliquin Reference Luca, Malhotra and Poliquin2020). This presents an unsolved puzzle. Why do salient mass shootings and a supportive public fail to instigate policy change? To solve this puzzle, scholars have examined whether mass shootings change electoral incentives and whether a lack of policy changes occurs in spite of or (perhaps) because of a lack of electoral pressure. In estimating the effects of these tragic shootings on elections, scholars have relied on panel data and difference-in-differences designs that exploit variation in shootings’ timing and location. Yet, despite using the same data sources, previous work reaches starkly different conclusions, with some finding mass shootings have strong electoral effects (García-Montoya, Arjona, and Lacombe Reference García-Montoya, Arjona and Lacombe2022; Yousaf Reference Yousaf2021) and others finding null effects (Hassell, Holbein, and Baldwin Reference Hassell, Holbein and Baldwin2020).

We show these conflicting findings come from the failure of some work to account for potential violations of the (essential) parallel trends assumption. Previous work documenting large effects of gun violence on electoral outcomes concludes so erroneously because shootings are more likely to happen in areas trending Democratic before shootings happened, whereas areas where shootings have not occurred were trending Republican.Footnote 1 (Note: if all the reader wants is a clear explanation of why research that does not account for parallel trends violations diverges in its conclusions from research that does account for trends, see Figures 2 [showing different trends for counties with and without shootings] and 4 [showing counties with shootings trending Democratic even before shootings]).

Figure 1. Differences in Previous Studies’ Estimated Effect of Mass Shootings on Election Outcomes Are Not Driven by Data Choices

Note: Estimates include county and year fixed effects (i.e., the TWFE estimator) with standard errors clustered at the county level and no covariates. The top panel shows effect estimates coding only the election immediately after a shooting occurs as having been treated; the bottom panel considers all post-shooting elections in counties with a shooting as treated. Coefficients, standard errors, and p-values are labeled for each coefficient. Takeaway: Naive TWFE estimators suggest mass shootings—regardless of the data/coding used—increase Democratic vote share in the county where shootings happens by 2.6–8.7 percentage points.

Figure 2. Trends in Presidential Vote in Counties with Mass Shootings Prior to Shootings, Compared to Trends in Counties without Shootings

Note: Pretreatment trends of Democratic vote share in counties where a shooting occurred (left panel) benchmarked to the trends in Democratic vote share found in counties where a shooting did not occur (right panel). Lighter lines show the patterns of individual counties; darker lines show the overall pattern for all counties. Takeaway: Counties that have shootings trended more Democratic even before the shootings occurred, whereas counties without a shooting trended slightly more Republican. Models that do not account for differential trends across counties will be biased.

Figure 3. The Effect of Shootings on Election Outcomes Many Years Before

Note: Effect of mass shootings on Democratic vote share in the years prior to when a shooting occurred. Treatment #1 is coded such that only elections with shooting are coded as treated; Treatment #2 is coded such that all elections after a shooting occurs in a county are coded as treated. All models’ standard errors are clustered at the county level. Takeaway: TWFE estimators without time-trends indicate shootings may have an effect up to and including 20 years prior to when a shooting occurred.

Figure 4. Event-Study Estimates Show that TWFE Fails to Account for Pretreatment Trends

Note: Event-study estimates with county and year fixed effects (GMAL’s data). Baseline election year is shown with a gray dotted line. Following prior work, we bin our extreme points (Baker, Larcker, and Wang Reference Baker, Larcker and Wang2022; Schmidheiny and Siegloch Reference Schmidheiny and Siegloch2019). Takeaway: Counties that have shootings trended more Democratic even before the shootings occurred. The increase that occurs after a shooting is entirely consistent with a general trend toward more Democratic election outcomes. Models that do not account for differential trends across counties will be biased.

Models accounting for violations of parallel trends provide no evidence that mass shootings cause large electoral change in the United States, and while chances of much smaller positive or negative effects cannot be entirely eliminated, almost all of these estimates are not statistically significant and are highly dependent on specific model specifications (see Figure 11). Sensitivity analyses embracing uncertainty around exact departures from parallel trends show these results are highly sensitive to minimal reasonable departures from parallel trends. Hence, the preponderance of evidence does not support conclusions that mass shootings have any large positive effect on Democratic vote shares.

These results are consistent whether we look at all mass shootings, school shootings, or just “rampage-style” school shootings.Footnote 2 (Note: if all readers want out of this article is a clear estimate of the effects [or more appropriately the lack thereof] of gun violence on Democratic vote shares, see Figures 5 [showing nonsignificant effects after controlling for unit time trends], 6 [showing nonsignificant effects after controlling for unit time trends in event-study designs], and 11 [showing the distribution of effects around zero].)

Figure 5. Effects of Mass Shootings on Elections after Absorbing County-Specific Trends

Note: Effect of mass shootings of various types once we account for differential trends. Within each panel, the first three estimates are using the GMAL coding of mass shootings and their data, the next comes from HHB, and the last comes from Yousaf. For cubic and quartic specifications, see Supplementary Figure S9. For effects where we code all post-shooting counties as being treated—not just counties and years with shootings—see Supplementary Figure S14. Takeaway: Once we account for differential trends across counties, the effects of mass shootings are all smaller and precisely estimated.

Figure 6. Event-Study Estimates of Shootings after Absorbing County-Specific Trends

Note: Event-study estimates from the HHB and GMAL data with county and year fixed effects and county-specific quadratic time trends. These use the method developed by Freyaldenhoven et al. (Reference Freyaldenhoven, Hansen, Pérez and Shapiro2021) to account for pre-trends in event-study designs. Analysis executed using the xtevent and xteventplot commands in STATA (Freyaldenhoven et al. Reference Freyaldenhoven, Hansen, Pérez and Shapiro2022). These commands, as a default, plot both the standard confidence intervals and those developed by Olea, Luis, and Plagborg-Møller (Reference Olea, Luis and Plagborg-Møller2019), which were developed for contexts with dynamic effects. The figure uses the same y-axis as Figure 4 for ease in comparing across the two. Takeaway: Once time trends are taken into account, the effect of shootings attenuates considerably.

Resolving discrepancies in these published findings also provides an opportunity to illustrate the critical importance of navigating pitfalls in difference-in-differences designs. The difference-in-differences design has recently proliferated, partially because of its simplicity and modest data requirements coinciding with a broader interest in causal inference and “credible” estimates which it provides (Angrist and Pischke Reference Angrist and Pischke2010). This has prompted a growing methodological literature covering the potential and pitfalls of this design (e.g., De Chaisemartin and d’Haultfoeuille Reference De Chaisemartin and d’Haultfoeuille2020; Kahn-Lang and Lang Reference Kahn-Lang and Lang2020; Roth et al. Reference Roth, Sant’Anna, Bilinski and Poe2022).

Although we strive first to answer the question of whether mass shootings affect election outcomes in the United States, we are also interested in narrowing the gap between theory and application of the difference-in-differences designs in political science. We do so by (1) outlining potential biases arising from (a) violations of parallel trends and (b) treatment effect heterogeneity; (2) highlighting the importance of researcher decisions related to specifying difference-in-differences models (e.g., treatment coding, the use of time trends and, if so, their functional form, and how [or at what level] to adjust standard errors); and (3) implementing them in an applied example.

This article provides a guide in the application of the difference-in-differences design and provides an answer to an important unresolved debate, shedding light on the political economy of gun violence in the United States and contributing to our understanding of what events spark electoral accountability.

DIFFERENCE-IN-DIFFERENCES AND THE TWO-WAY FIXED EFFECTS ESTIMATOR

We think it is important to first explain the predominant difference-in-differences approach and the logic behind the pitfalls that exist. Difference-in-differences designs routinely rely on two-way fixed effects estimator (TWFE). With TWFE, the outcome of interest is regressed on time and unit (often geographic) fixed effects, along with the treatment status. The TWFE controls for factors remaining constant within years (e.g., nationwide economic conditions) and factors varying across spaces (e.g., stable local culture).

Original difference-in-differences designs used these identification strategies in largely exogenous interventions implemented in a single time period. This design constitutes a two-group (treated and not treated) and two-period (pre and post) design and “[the difference-in-differences estimator] is equal to the treatment coefficient in a TWFE regression with group and period fixed effects” (De Chaisemartin and D’Haultfoeuille Reference De Chaisemartin and D’Haultfoeuille2023, C3).

Importantly, this design rests on the parallel trends assumption or the assumption that without treatment “that the average outcome among the treated and comparison populations would have followed ‘parallel trends’ in the absence of treatment.”Footnote 3 As discussed below, there are multiple ways to evaluate and adjust for violations of this crucial assumption.Footnote 4

DIFFERENCES IN FINDINGS ON THE ELECTORAL EFFECTS OF MASS SHOOTINGS

In an article published at the American Political Science Review (APSR), Hassell, Holbein, and Baldwin (Reference Hassell, Holbein and Baldwin2020) (hereafter HHB) estimate the effect of school shootings on voter turnout and election outcomes at federal, state, and local levels. Using various modeling strategies, HHB find school shootings—regardless of the number of victims—have precisely estimated null effects on vote shares.Footnote 5 In contrast, in a later article in APSR, García-Montoya, Arjona, and Lacombe (Reference García-Montoya, Arjona and Lacombe2022) (hereafter GMAL) focus on the effect of “rampage-style” school shootings. Using a TWFE, they emphasize (in the abstract and throughout the manuscript) that these shootings increase Democratic vote share by around 5 percentage points in the local community.Footnote 6 In work published in the Journal of the European Economic Association, Yousaf (Reference Yousaf2021) uses a TWFE showing all mass shootings—not restricted to school shootings—decrease Republican presidential vote share by 2–6 percentage points locally.Footnote 7

Going one step further, there are model specifications that are similar to those that GMAL and Yousaf run on their data and are theoretically justified but that are not run in these original papers. Some of these plausible alternative specifications (albeit, these are not explored by GMAL or Yousaf) suggest effects as large as $ \approx $ 8.7 percentage points the election following a mass shooting. In addition, some event-study models using TWFE specifications (albeit, these are not explored by GMAL or Yousaf) suggest effects as large as 13 percentage points a full 28 years after a mass shooting.

Ultimately, GMAL’s and Yousaf’s conclusions differ with HHB’s—with the former suggesting large statistically detectable meaningful effects of mass shootings in partisan vote shares in local communities. Our work shows results from TWFE models suggesting these large effects are not robust because they fail to fully account for violations of the critical parallel trends assumption.

DATA/METHODS

All papers in this literature use a common dataset—Dave Leip’s Atlas of U.S. Elections—which reports county-level vote shares.Footnote 8 We focus our examination on key differences in previous work examining the electoral effects of gun violence.Footnote 9 In estimating difference-in-differences designs, one decision in the hands of the researcher is what counts as treatment. In our case, each study uses slightly different shootings as treatments (see Supplementary Table S1). However, despite previous claims (GMAL, 821–3), differences in data choices and coding are ultimately not what drives divergent findings.

Below, we are interested not only in the statistical significance of the effects but also in their magnitude. While adjudicating effect size is always somewhat in the eye of the beholder, we use several tools to quantify the size of our observed effects. First, we benchmark estimates to other similar geographic-based treatments. Second, we use equivalence testing to see what effects we are able to rule out (Hartman and Hidalgo Reference Hartman and Hidalgo2018). Finally, though not fully capturing the scope of effects, we also note when effects are not statistically significant.

TWFE ESTIMATOR IN OUR EMPIRICAL CASE

As detailed previously, the most common approach to estimating difference-in-differences effects when treatment varies over time and space is the TWFE. This approach is followed by GMAL and Yousaf, who emphasize results using county and year fixed effects.Footnote 10 We replicate this approach with the data provided by HHB although this is not their primary estimator.Footnote 11 The TWFE is specified in Equation 1, where $ {Y}_{ct} $ represents the Democratic vote share in a county (c) and election period (t), $ {\phi}_c $ represents a county fixed effect, $ {\lambda}_t $ represents a year fixed effect, $ {\epsilon}_{ct} $ represents the error term, and $ {D}_{ct} $ denotes the treatment (i.e., whether a county (c) in a given election period (t) is exposed to a mass shooting). $ \beta $ is the effect mass shootings on Democratic vote share:Footnote 12

(1) $$ \begin{array}{rl}{Y}_{ct}={\phi}_c+{\lambda}_t+\beta *{D}_{ct}+{\epsilon}_{ct}.& \end{array} $$

Another researcher decision point is how to code the exact nature of the treatment.Footnote 13 One possibility is to code units exposed to treatment in a given period as treated and all other observations—pre- and posttreatment in eventually treated units and never-treated units—as the control group. In the mass shooting example, this approach codes all counties with a mass shooting in a given electoral cycle as treated, but county-level observations before and after that electoral cycle (along with those who never have a shooting) as untreated, allowing treatment in- and out-switchers. This approach assumes effects of mass shootings are constrained to the immediate electoral cycle. Alternatively, another approach is to code treatment, so all observations in treated units posttreatment are coded as treated. In our example, this approach codes all counties with mass shootings in an election cycle and following election cycles as treated, and all counties before—along with counties never having shootings—as untreated. This means there are no out-switchers. This approach allows mass shootings to have longer effects, changing the electoral environment both when they occur and afterward.Footnote 14

The choice between the two approaches is a researcher decision that should be motivated by theory. Here, given a lack of strong expectations about mass shootings temporal effects, we use both approaches. (We complement these approaches with an event-study design described below, explicitly modeling effects in periods before and after shootings with lags and leads.)

TWFE Estimates of Shootings’ Effect on Vote Shares

We start by showing that different conclusions across studies on the electoral effects of shootings are not driven by data choices. Figure 1 removes differences in methodological approaches in previous studies and shows the effects of TWFE models (GMAL’s and Yousaf’s approaches) using the data from all of the studies. Figure 1 also splits the results by treatment coding approaches outlined above.Footnote 15

As shown in Figure 1, TWFE specifications consistently produce substantive positive statistically significant effects regardless of the time frame, treatment codings—be they “rampage-style” school shootings (GMAL), school shootings (HHB), or all mass shootings (Yousaf)—or how long treatments apply. With GMAL’s data, we find—like GMAL did—that “rampage-style” shootings correlate with increases in Democratic vote share. This estimate (while varying by specifications) is 5.5 percentage points in the TWFE estimator for the first treatment coding (i.e., panel a) and is 8.7 percentage points in the second treatment coding (i.e., panel b). Both estimates are highly significant ( $ p<0.01 $ ). The effects for HHB and Yousaf are very similar. Simply, when using the same model choices, the effects of shootings of different types are consistently sizeable. In other words, previous differences in conclusions across studies of the electoral effects of gun violence are not due to choices about which shootings count as treatment. In short, TWFE estimates, regardless of the coding of shootings, indicate significant and substantively meaningful positive effects of shootings on Democratic vote share.

We pause to discuss effect magnitude. Upwards of an 8.7 percentage point shift in Democratic vote share is large—as are many of the other estimates. As GMAL note, these effects represent “a remarkable shift in an age of partisan polarization and close presidential elections” (GMAL, 809). We see how large these effects are by benchmarking them to other studies using county-level vote shares and difference-in-differences designs. For example, Sides, Vavreck, and Warshaw (Reference Sides, Vavreck and Warshaw2022, 709) estimate a six-standard-deviation shift in relative television advertising produces a 0.5-point change in two-party vote share. Hence, if we believe these results travel, GMAL’s simple TWFE estimates indicate one school shooting has an effect on Democratic vote share equivalent to a shift of approximately 66–104 standard deviations in relative advertising. Using an economic comparison—the most common of retrospective voting treatments—Healy and Lenz (Reference Healy and Lenz2017, 1423) show a “1 percentage point increase in mortgage delinquencies increases Democratic vote share by 0.33 percentage points.” Thus, the effect of “rampage-style” shootings is roughly the equivalent to a 16.7–26.4 percentage point increase in mortgage delinquencies; or moving from a world with no delinquencies to one where about one-fifth of residents are at risk of losing their homes.

In short, TWFE models suggests gun violence—regardless of how shootings are coded—fundamentally reshapes electoral results in the local communities in which they occur. Is this sizable relationship causal and robust? Recent methodological developments provide us a guide to answer this question.

ADDRESSING ISSUES WITH TWO-WAY FIXED EFFECTS ESTIMATORS

Recent research has shown that simple TWFE models can be problematic for important reasons, including:

  1. 1. violations of the parallel trends assumption (e.g., Freyaldenhoven et al. Reference Freyaldenhoven, Hansen, Pérez and Shapiro2021; Liu, Wang, and Xu Reference Liu, Wang and Xu2024; Rambachan and Roth Reference Rambachan and Roth2021) and

  2. 2. mistaken inferences derived from heterogeneity in treatment effects (e.g., Goodman-Bacon Reference Goodman-Bacon2021; Sun and Abraham Reference Sun and Abraham2021).

We discuss these issues in order and apply solutions articulated in the literature using the example of mass shootings’ seeming effects on Democratic vote share.

ASSESSING AND ADDRESSING PARALLEL-TRENDS VIOLATIONS

A core assumption to the difference-in-differences design is the parallel trends assumption. The parallel trends assumption asserts outcomes of interest from pre- to posttreatment would have moved in parallel across treated and untreated groups if not for treatment. If parallel-trends assumptions are violated, estimated effects are biased. There are several ways to assess the potential for differential pretreatment trends.

Checking for Visual Evidence of Differential Pretreatment Trends

Because of the fundamental problem of causal inference, we do not observe counter-factual worlds where the treated and untreated groups are exposed to opposite conditions. Hence, no singular test can prove parallel trends is satisfied; however, treated and untreated units not moving together before treatment exposure, indicates potential issues (De Chaisemartin and D’Haultfoeuille Reference De Chaisemartin and D’Haultfoeuille2023; Marcus and Sant’Anna Reference Marcus and Sant’Anna2021).Footnote 16 An appropriate first step is to visually inspect patterns in aggregate-level data to see whether, prior to treatment, treatment and control areas are trending in different directions.Footnote 17

Figure 2 examines differential pretreatment trends separating counties into two bins; panel (a) contains counties with a shooting excluding all post-treatment observations, and panel (b) contains all counties without shootings.Footnote 18 , Footnote 19 Figure 2 illuminates what TWFE models absorb and do not absorb. County fixed effects adjust for differences in Democratic vote share across counties. Year fixed effects account for differences across years. However, TWFE models do not account for the possibility counties’ Democratic vote shares change at different rates over time, a significant problem in the context of school shootings.

As Figure 2 shows, shootings happen in communities—proceeding and unrelated to shootings themselves—trending more Democratic relative to other locales. This is likely because mass shootings occur disproportionately in growing populations and have increased over time (Musu-Gillette et al. Reference Musu-Gillette, Zhang, Wang, Kemp, Diliberti and Oudekerk2018; U.S. Government Accountability Office 2020) at the same time American politics has realigned with these same more populated areas becoming more Democratic (DeSilver Reference DeSilver2016). In general, researchers should take care when demographic/political changes predating treatment aligns with short-term treatment exposure.

This (coincidental) pretreatment trend separation becomes especially prevalent after 2004. This is particularly problematic because GMAL explicitly note shootings effects before 2004 are essentially null (or negative), but in 2004 the positive effects on Democratic vote share increase (GMAL, Figure 7). Figure 2 indicates this is the exact time when parallel trends assumptions become particularly tenuous. Two-way fixed effects models do not absorb these trends and, as such, are likely biased estimates.

Checking for Pretreatment Effects with the Model Specifications Used

While Figure 2 provides visual evidence of differences in pretreatment trends, it is not dispositive. Graphical representations can differ and changes in formatting can minimize or exacerbate the appearance of differential trends leading researchers to draw different conclusions.Footnote 20 Hence, the next check to assess TWFE design validity which should be standard practice is whether this model suggests impacts prior to treatment (Grimmer et al. Reference Grimmer, Hersh, Meredith, Mummolo and Nall2018). In our case, this placebo test is informative as shootings—something that people cannot precisely anticipate—should not affect vote shares prior to their occurrence. If there are effects, it suggests TWFE estimates are likely not causal (Angrist and Pischke Reference Angrist and Pischke2008; Hansen and Bowers Reference Hansen and Bowers2008).

Specifically, we run the specification listed in Equation 2. Equation 2 is the same as Equation 1, except for the outcome variable. Instead of estimating shootings’ effects ( $ {D}_{ct} $ ) in subsequent elections ( $ {Y}_{ct} $ ), we substitute a lagged outcome variable ( $ {Y}_{ct-k} $ ). Here, k corresponds to the number of lagged periods included. We include seven lagged periods in our models as the GMAL panel is sufficiently long. However, power considerations may influence the number of lags used. We recommend initially looking for effects in one period lag, then looking how far back one can estimate precise-enough specifications:Footnote 21

(2) $$ \begin{array}{rl}{Y}_{ct-k}={\phi}_c+{\lambda}_t+\beta *{D}_{ct}+{\epsilon}_{ct}.& \end{array} $$

Panels a and b (the top section) of Figure 3 (we discuss panels c–f later) show the TWFE models for various shootings on lagged measures of Democratic vote share and show there is substantial imbalance in lagged outcomes.Footnote 22 , Footnote 23 We start on the left of each panel, with the presidential election prior to the shooting and work up to seven presidential elections (28 years) before shootings occurred.Footnote 24 Effects vary by specification, but range between 2 and 7 percentage points, with most highly significant. This analysis indicates mass shootings have a significant and substantive effect on Democratic vote shares up to 20 years prior to a shooting. Simply, the TWFE does not recover balance prior to shootings, regardless of data used.

There is little reason—theoretically or empirically documented—to suspect shootings should have anticipatory effects on vote shares, given these events are unexpected where they occur. There is, however, the possibility pretreatment effects show up where there is not bias if treatment in one period ( $ {D}_{ct} $ ) is highly correlated with treatment in prior periods ( $ {D}_{c,t-k} $ ). In that case, the coefficient on $ {D}_{ct} $ may show an effect if there is an effect of $ {D}_{c,t-k} $ on $ {Y}_{c,t-k} $ . In such cases, pretreatment effects could emerge even without errors in the research design. Therefore, we recommend also modeling effects of lagged and leaded treatment using an event-study design.

Checking for Pretreatment Trends with Event-Study Designs

An event-study design traces effects before and after treatment and provides another way to see pretreatment imbalances (Armitage Reference Armitage1995; Binder Reference Binder1998). An event-study is an increasingly common difference-in-differences model, given its less restrictive and more transparent modeling assumptions, but still relatively rare in political science. An event-study (usually) uses TWFE but also includes lagged and lead treatment variables as shown in Equation 3 below. We list treatment in a given county (c) and year (t), lagged or leaded by the corresponding periods since treatment (k). For simplicity, Equation 3 shows the event-study model for only one pretreatment period ( $ k-2 $ ), the period when treatment occurs (k), and one period after treatment occurs ( $ k+1 $ ). The baseline is the period before treatment occurs ( $ k-1 $ ) (Armitage Reference Armitage1995; Binder Reference Binder1998):

(3) $$ \begin{array}{rl}{Y}_{ct}={\phi}_c+{\lambda}_t+{\beta}_{-2}*{D}_{ct,k-2}+{\beta}_0*{D}_{ct,k}+{\beta}_1*{D}_{ct,k+1}+{\epsilon}_{ct}.& \end{array} $$

Figure 4 shows nine preelection treatments and eight posttreatment periods included using the GMAL data. The right of Figure 4 (right of the gray vertical line) shows an immediate, significant, and substantive jump in Democratic vote share the election year following a shooting and grows after the event occurs, having an increasingly larger effect on Democratic vote share (10–13 points), and allows us to rule out smaller effects using equivalence testing. While not completely impossible, the long-lasting and growing effect remains theoretically unexplained.

However, by looking to the left of the baseline period (left of the gray dotted vertical line), Figure 4 also illuminates the “effect” is unlikely to be causal. If the TWFE models were causal, these coefficients should not be significantly/substantively distinct from zero. Instead, Figure 4 shows that relative to one election prior to a shooting, prior election years see less Democratic support and this underperformance increases further back in time. Simply, vote shares trend more Democratic before shootings in counties where shootings occur (relative to counties without shootings). Elections after a shooting are just a continuation—indeed, the points represent an almost perfect linear function—further evidence that TWFE estimators are biased in this case. We recommend parameterizing models as an event-study become standard in difference-in-differences applications.

Controlling for Any Differential Unit-Specific Time Trends

Facing potential parallel-trends violations, one potential remedy is adjusting for factors—observed or unobserved—leading to pretreatment imbalances. In this case, visual inspection (see Figure 2) reveals treated and untreated units have different pre-trends. A solution is to include unit-specific time trends (Angrist and Pischke Reference Angrist and Pischke2008; Reference Angrist and Pischke2010; Wing, Simon, and Bello-Gomez Reference Wing, Simon and Bello-Gomez2018), making the identifying assumption deviation from county-year trends. Identification comes from sharp deviations from otherwise smooth unit-specific trends corresponding to the following equation:

(4) $$ \begin{array}{rl}{Y}_{ct}={\phi}_c+{\lambda}_t+{\gamma}_c*t+\beta *{D}_{ct}+{\epsilon}_{ct}.& \end{array} $$

While Equation 4 includes linear county-specific time trends ( $ {\gamma}_c*t $ ), the functional form is a potentially influential decision in the hands of a researcher.Footnote 25 As a result, we run many model specifications—all taking slightly different tacts to adjusting for differential pre-trends. The next section shows various other approaches, including methods recently developed by Liu, Wang, and Xu (Reference Liu, Wang and Xu2024), Freyaldenhoven et al. (Reference Freyaldenhoven, Hansen, Pérez and Shapiro2021), and Rambachan and Roth (Reference Rambachan and Roth2021). We recommend scholars run robustness checks across several parameterizations of the unit-specific trends, acknowledging higher-order unit-specific trends could face a bias-variance tradeoff, especially in smaller datasets.

Figure 3c–f shows the pretreatment effect models with linear and quadratic trends (adding $ {\gamma}_c*{t}^2 $ to Equation 4). For cubic and quartic county-specific trends, see Supplementary Figures S9 and S10. In contrast to the TWFE (Figure 3a,b), all models with county-specific trends are balanced pre-treatment. Importantly, these null effects (especially in proximate periods) are not driven by standard error inflation ruling out even modest pretreatment differences using equivalence testing (effects outside the $ -0.99 $ to 0.57 percentage point range).

Figure 5 shows estimates for models including linear and quadratic county-specific time trends on the posttreatment outcomes with only counties with shootings in that year coded as treated.Footnote 26 As Figure 5 shows, once we make this necessary adjustment to correct for the violation of the parallel trends assumption, the effects of mass shootings of all types attenuate heavily. All effect estimates are much smaller and virtually all are not significant.

Specifically, Figure 5a,b—the estimates for “rampage-style” school shootings (i.e., GMAL’s treatment)—indicates a 0.7 percentage point increase in Democratic vote share, an effect that is not statistically significant and 7.9 times (i.e., 790%) smaller than the original TWFE effectively ruling out effects as large as the TWFE with a high degree of confidence.Footnote 27 , Footnote 28 We can effectively rule out effects of “rampage-style” shootings (GMAL’s treatment) larger than 1.5 percentage points and smaller than $ -0.1 $ percentage points, effects of all mass shootings (Yousaf’s treatment) larger than 1.2 percentage points and smaller than $ -1.6 $ percentage points.Footnote 29 Effects of all school shootings (HHB’s treatment) are also much smaller and not significant. Simply, there is no evidence of large effects documented in previous work finding an effect, and no consistent evidence for positive (or negative) effects statistically distinguishable from zero regardless of the data used.Footnote 30 While statistically significant effects infrequently show up in Figure 5, they are not robust. Notably, if we add higher-order polynomials—as in Supplementary Figure S9—no effects are significant. This—along with further checks below—emphasizes the lack of support for the conclusion that shootings have significant, systematic, or large effects on Democratic vote shares.Footnote 31

Using an event-study design that accounts for the differential pre-treatment trends by adding leads and lags of the treatment values bolsters conclusions as effects are even smaller and more precise. Figure 6 uses Freyaldenhoven et al. (Reference Freyaldenhoven, Hansen, Pérez and Shapiro2021) methods displaying event-study designs and accounting for pre-trends in event-study designs.Footnote 32 , Footnote 33 Figure 6 uses the same y-axis as Figure 4 for ease in comparison. Adding trends attenuates pre-treatment imbalances. So too, however, are any large post-treatment differences as the immediate effect of “rampage-style” school shootings is a mere 0.8 percentage point bump for Democrats. This effect is barely statistically significant using the linear trends models (p=0.048) but still allows us to confidently rule out large effects; we can rule out effects larger than 1.67 percentage points using equivalence testing, nowhere near the size of GMAL’s 5 percentage point estimate emphasized throughout their text. Using the quadratic trends model, the effect is only marginally significant at the 10% level (p<0.066) and we can rule out effects larger than 1.71 percentage points.Footnote 34 These effects appear to be the upper bound produced from this method.

If we use Clarke and Tapia-Schythe’s (Reference Clarke and Tapia-Schythe2021) approach to estimating the event-study with trends and the corresponding eventdd command in STATA, we get negative effects on Democratic vote shares (albeit statistically indistinguishable from zero). With this slightly different approach, the effects in elections immediately following shootings are $ -0.13 $ percentage points ( $ p=0.898 $ ; 95% CI: [−2.05, 1.80]).Footnote 35 Other shooting codings are similarly very small and insignificant. Moreover, testing the robustness of these effects to other pretreatment periods—as the approach designed by Freyaldenhoven et al. (Reference Freyaldenhoven, Hansen, Pérez and Shapiro2021) allows—effects are even smaller and even less suggestive of an effect (see Supplementary Figures S2–S5). Benchmarked to the two-period lag trend, the estimate for elections immediately after a shooting using linear county trends is 0.47 percentage points ( $ p=0.324 $ ; 95% CI: [−0.5, 1.4]). Estimates for elections after a shooting for the quadratic county trends model is 0.5 percentage points ( $ p=0.359 $ ; 95% CI: [−0.5, 1.5]). Moreover, none of the large longer-term effects remain.

In short, in event-studies adjusting for unit-specific trends, there is little evidence for the sizeable effects previously suggested. In fact, there is little evidence for any significant effect. The occasional effect crossing the $ p<0.05 $ threshold are not robust to reasonable model variations, such as the baseline comparison points one uses. Moreover, while some specifications cannot fully exclude some much smaller but non-negligible positive effects, most 95% confidence intervals also cannot rule out non-negligible negative effects.

Additional Checks Addressing Violations of Parallel Trends Assumptions

Including unit-specific time trends (see above) is not the only solution to parallel-trends violations; indeed, scholars may desire more flexible solutions. Recent advances have suggested many alternative solutions to potential parallel-trends violations or unobserved time-varying confounders. Unfortunately, these approaches have not yet been benchmarked to one another, let alone compared in different data contexts. The standard approach is to develop a new estimator and assert it applies in all contexts. We are currently unaware of work testing these methods head-to-head, let alone providing recommendations regarding which methods to apply in different contexts. As such, we recommend scholars run a variety of specifications, as we do below, identifying general patterns. We recommend that scholars implement, at minimum, checks suggested by Liu, Wang, and Xu (Reference Liu, Wang and Xu2024) and Rambachan and Roth (Reference Rambachan and Roth2021) outlined below.

Building on research exploring factor-augmented models for causal identification (e.g., Bai and Ng Reference Bai and Ng2021; Xu Reference Xu2023), Liu, Wang, and Xu (Reference Liu, Wang and Xu2024) develop procedures—including what they call the fixed effects counterfactual estimator, the interactive fixed effects counterfactual estimator, and the matrix completion estimator—to “estimate the average treatment effect on the treated by directly imputing counterfactual outcomes for treated observations” (1).Footnote 36 Using simulations, Liu, Wang, and Xu (Reference Liu, Wang and Xu2024) show that the interactive fixed effects counterfactual estimator provides more reliable causal estimates than conventional TWFE models when unobserved time-varying confounders exist. The interactive fixed effects counterfactual estimator can be applied with the package panelView, which is available in both Stata and R (Mou, Liu, and Yiqing Reference Mou, Liu and Yiqing2022b).

Figure 7 applies Liu, Wang, and Xu’s (Reference Liu, Wang and Xu2024) approach using the GMAL data.Footnote 37 Panel a shows the TWFE and panels b–d show interactive fixed effects counterfactual estimators. In the TWFE model, there are pretreatment imbalances and a general overall upward trend—as Figure 5 indicated previously. Again, this suggests the TWFE picks up on a general pro-Democratic trend in pre-treatment periods. However, after adjusting for pretreatment differences using Liu, Wang, and Xu’s (Reference Liu, Wang and Xu2024) approach to address pretreatment imbalances, there is virtually no evidence shootings substantially or significantly affect vote shares in subsequent elections. Nor is this for lack of statistical power and we can rule out larger effects using equivalence testing. Moreover, any (much smaller) effects that appear intermittently are not robust to reasonable model variations in the realm of researcher decision-making.

Figure 7. Liu, Wang, and Xu (Reference Liu, Wang and Xu2024) Interactive Fixed Effects Counterfactual Estimator

Note: The interactive fixed effects counterfactual estimator developed by Liu, Wang, and Xu (Reference Liu, Wang and Xu2024) using GMAL’s data. Panel a shows the TWFE estimated by Liu, Wang, and Xu’s (Reference Liu, Wang and Xu2024) FECT package; it is analogous to Figure 5, but their procedure estimates slight differences—for example, in the number of pre- and posttreatment periods. In panel b, the number of factors (r) is set to 3—that chosen by cross-validation and the degree of the polynomial is set to 4. In the bottom row, r is set to 1 in both panels and degree 2 in panel c and 4 in panel d. For other variations, see the Supplementary Material. Takeaway: The upward trend in the TWFE model (i.e., panel a) is indicative of violation of the parallel trends assumption. In the interactive fixed effects models, there is no evidence of the substantial effects shown in more simplistic model specifications that do not account for potential violations of the parallel trends assumption.

Rambachan and Roth (Reference Rambachan and Roth2021) propose another solution using a sensitivity analysis approach for potential violations of the parallel trends assumption allowing researchers to avoid arbitrarily choosing a parametric model.Footnote 38 This approach is particularly useful as researchers often struggle to know the functional form of the underlying system.

This sensitivity analysis can be formalized in several ways. For example, researchers can see how robust their effects are to varying departures from differential trends evolving smoothly over time. This may be especially useful when concerned “about confounding from secular trends … evolv(ing) smoothly over time” (Rambachan and Roth Reference Rambachan and Roth2021, 13)—as we are in this case. This sensitivity test is “done by bounding the extent to which the slope may change across consecutive periods” (Rambachan and Roth Reference Rambachan and Roth2021, 12).Footnote 39 They call this the $ SD $ or “second derivative” or “second differences” approach.Footnote 40 Using Rambachan and Roth’s (Reference Rambachan and Roth2021) general approach, conclusions do not depend on arbitrary model specification choices. In essence, this approach “show[s] what causal conclusions can be drawn under various restrictions on the possible violations of the parallel-trends assumption” (Rambachan and Roth Reference Rambachan and Roth2021, 1). This approach is implementable through the HonestDID package in R and STATA (Rambachan and Roth Reference Rambachan and Roth2021; Rambachan, Roth, and Bravo Reference Rambachan, Roth and Bravo2021).

Rambachan and Roth (Reference Rambachan and Roth2021, 28) note “it is natural to report both the sensitivity of the researcher’s causal conclusion to the choice of this parameter and the ‘breakdown’ parameter value at which particular hypotheses of interest can no longer be rejected.” Figure 8 shows Rambachan and Roth’s sensitivity approach using GMAL’s and HHB’s data. Figure 8 uses the $ SD $ approach, plotting robust confidence sets for the treatment effect in the mass shooting case for different values of the parameter M. The confidence sets show that the effect of mass shootings on Democratic vote share is only positive and significant in the coefficient on the far left—the original estimate not allowing for any violations of parallel trends—indicating effects of mass shootings are highly sensitive to any minor departures from parallel trends. The robust confidence intervals include zero when allowing for linear violations of parallel trends ( $ M=0 $ ), and become even wider allowing for nonlinear violations of parallel trends ( $ M>0 $ ). Such a low breakdown suggests any meaningful departure of the slope changing between consecutive periods, would cause the observed effects in GMAL’s data to not be significant. Overall, these results indicate effects of shootings (using the GMAL data) are highly sensitive to even minor parallel-trends violations.

Figure 8. Implementing Rambachan and Roth’s Sensitivity Analysis in the Shooting Example

Note: Results from the sensitivity analysis suggested by Rambachan and Roth (Reference Rambachan and Roth2021) using GMAL’s data—that is, testing for effect sensitivity across $ {\triangle}^{SD}(M) $ . The models incorporate information from three elections prior to treatment and five post-treatment periods. Takeaway: The results show the effects of shootings on vote shares are highly sensitive and do not hold with even minor deviations from parallel trends.

Summarizing Tests for Violations of Parallel Trends

Testing for potential violations of parallel trends is often not easy or straightforward because of the fundamental problem of causal inference—that is, the inability to observe what would have happened to treated groups without treatment. But even when considering only tests of pre-trends, there are challenges. For example, in choosing methods to address this core issue, one must consider the nature of the data available, the statistical power, and numerical degrees of freedom (Bilinski and Hatfield Reference Bilinski and Hatfield2018; Roth Reference Roth2022). We are not arguing every case employ unit-specific trends. What we are arguing is that all researchers should diagnose and address potential violations of this core assumption. How they do so—with the many tools at their disposal that we have outlined above—is less important than that they do so in a way that is transparent, thorough, and appropriate to the applied case.

In the case of mass shootings, this much is clear: once we make adjustments for clearly differential pre-trends, the evidence for any effect gets much more muddled than previous studies have suggested. Effects are substantively smaller than what simple TWFE models suggest. There is no clear evidence of durable lasting effects. Moreover, depending on the dataset one uses and the way treatment is coded, effects can be marginally positive, negative, or (approximately) zero, are very rarely significant, and highly sensitive to any meaningful departures from strictly parallel trends.

DIAGNOSING AND ADDRESSING TREATMENT EFFECT HETEROGENEITY/REMOVING CONTAMINATED COMPARISONS

Recent research also indicates issues with variations in treatment with heterogeneity of treatment effects. For example, Goodman-Bacon (Reference Goodman-Bacon2021) shows any TWFE estimate with variations in treatment timing can be decomposed into an average of all possible $ 2\times 2 $ difference-in-differences estimates constructed from the panel data set weighted by group sizes and variance in treatment (Goodman-Bacon Reference Goodman-Bacon2019). If there are time-varying treatment effects, they can produce biased estimates (Goodman-Bacon Reference Goodman-Bacon2019, 3).Footnote 41 To diagnose this potential for bias, Goodman-Bacon (Reference Goodman-Bacon2019) allows for decomposing the $ 2\times 2 $ difference-in-differences estimates using the bacondecomp package in R and STATA (Goodman-Bacon Reference Goodman-Bacon2021; Goodman-Bacon, Goldring, and Nichols Reference Goodman-Bacon, Goldring and Nichols2019).Footnote 42 As stated in the package’s STATA help file:

[The decomposition] by default produces a graph for all comparisons and shows up to three types of two-group/two period comparisons, which differ by control group: (1) Timing groups, or groups whose treatment stated at different times can serve as each other’s controls groups in two ways: those treated later serves as the control group for an earlier treatment group and those treated earlier serve as the control group for the later group; (2) Always treated, a group treated prior to the start of the analysis serves as the control group; and (3) Never treated, a group which never receives the treatment serves as the control group.

Figure 9 provides this decomposition.Footnote 43 As shown in both the GMAL and HHB data, the TWFE is a composite of $ 2\times 2 $ ’s eliciting large negative and positive effects. Effects can be very different depending on the $ 2\times 2 $ ’s included in the estimate (as indicated by the spread of estimates across the y-axis). Second, the TWFE is a weighted composite highly influenced by several comparisons of always treated versus timing. However, many of the $ 2\times 2 $ estimates have similar weight—as noted by the cluster of estimates on the left side of the graph. Overall, to our eye, there appears to be no clear evidence our effects are driven by treatments of various types.

Figure 9. Illustration of the Goodman–Bacon Decomposition of the TWFE Models

Note: This figure shows the results from the Bacon decomposition for the TWFE models. The figure also shows all of the possible $ 2\times 2 $ difference-in-differences (DiD) estimate, with their weights for the ATE on the x-axis and the effect size on the y-axis. The horizontal line shows the overall DiD estimate.

We think it is important to note that the mass shooting example is not an ideal application to show the value of Goodman–Bacon decomposition as (at present) the Goodman–Bacon decomposition only decomposes the TWFE and does not decompose more sophisticated models implemented accounting for parallel-trends violations. Nevertheless, this may not always be true in all contexts, and scholars should examine this decomposition to illuminate whether effects are driven by specific comparisons and if the bias Goodman-Bacon (Reference Goodman-Bacon2021) discusses is present.

There are several proposed solutions to problems arising with heterogeneous treatment effects (Borusyak, Jaravel, and Spiess Reference Borusyak, Jaravel and Spiess2021; Callaway and Sant’Anna Reference Callaway and Sant’Anna2021; De Chaisemartin and d’Haultfoeuille Reference De Chaisemartin and d’Haultfoeuille2020; de Chaisemartin, D’Haultfoeuille, and Guyonvarch Reference de Chaisemartin, D’Haultfoeuille and Guyonvarch2019; Zhang Reference Zhang2022). Some allow the inclusion of additional covariates including unit-specific trends. One approach that does is Sun and Abraham (Reference Sun and Abraham2021) implemented using the eventstudyinteract package in STATA and fixest package in R (Berge, Krantz, and McDermott Reference Berge, Krantz and McDermott2022; Sun Reference Sun2021).Footnote 44 , Footnote 45 This approach “estimates the shares of cohort as weights.” In our case, implementing Sun and Abraham’s solution with a simple TWFE (with no time trends and thus not addressing parallel-trends violations) substantially reduces effect estimates (Figure 10). At first, these changes look modest with event-study estimates going from 3 percentage points ( $ p=0.000 $ ) in the naive models to 2.4 percentage points ( $ p=0.001 $ ) in the Sun and Abraham adjusted models in the first posttreatment period. However, in the second and following treatment periods, effects that were large (10–13 percentage points) heavily attenuate and even become negative (1.1 [ $ p=0.19 $ ], $ -0.4 $ [ $ p=0.75 $ ], $ -2.2 $ [ $ p=0.12 $ ], and $ -1.5 $ [ $ p=0.18 $ ] in posttreatment elections 1–4, respectively). Adding unit-specific trends to Sun and Abraham’s estimator (panel a) makes effects even smaller. We go from 2.4 percentage points [ $ p=0.001 $ ] in the TWFE to roughly half (linear trends [ $ p=0.031 $ ]; quadratic [ $ p=0.051 $ ]; or cubic trends [ $ p=0.085 $ ]). (Even in the cubic model, the standard error remains modest in size—being 0.8 percentage points.) Moreover, the long-run effect of 10–13 percentage points is not present. None of the effects are present in the HHB data (panel b). This suggests that effect heterogeneity plays some role. Once adjusted for, long-term effects attenuate substantially and short-term effects become much smaller and flimsier to reasonable alternative specifications (e.g., the coding of treatment or the functional form of unit-specific trends).

Figure 10. Sun and Abraham (Reference Sun and Abraham2021) Approach for Estimating Dynamic Treatment Effects in Event Studies with Heterogeneous Treatment Effects

Note: Results from the clean comparisons suggested by Sun and Abraham (Reference Sun and Abraham2021) using the GMAL and HHB data. Models include quadratic county-specific time trends to address potential violations of the parallel trends assumption in the TWFE. Takeaway: Clean comparison effects with trends show no sign of a sizable and durable effect on Democratic vote shares shown in the TWFE nor in the simple event-study plot (see Figure 5).

More importantly, for reasons we outline below, it is unwise to overemphasize one model specification. Combining the evidence from all of the various approaches taking into account potential contamination from treatment effect heterogeneity, the best evidence suggests that in the case of shootings and electoral vote shares, (1) violations of parallel trends loom large, (2) effect heterogeneity may play a modest role, and (3) there is no sign of the sizable and durable effect, but perhaps a much smaller effect—and one that is not clearly distinct from zero across almost all reasonable model specifications (and is negative in some specifications).

OTHER POTENTIAL ISSUES RAISED IN THE LITERATURE ABOUT EXECUTING DIFFERENCE-IN-DIFFERENCES DESIGNS

A few final words of guidance and caution remain. First, in all empirical checks, it is important not to forget theory. For example, we focus on whether shootings change Democratic vote shares in counties where they occur. We have done so because whether an effect occurs there is the central dispute. However, there are other potential effects. Perhaps shootings have spillover effects—with effects in adjacent counties—or as a function of the distance from a shooting, the time since a shooting, or the intensity of the shooting (e.g., the number of deaths/injuries). Alternatively, perhaps mass shootings have national effects. The best evidence currently suggests that none of these things occur (HHB).Footnote 46 However, scholars should remember there are often multiple ways of conceptualizing treatment exposure.

Second, it may, at times, be useful to unpack treatment effects at a more granular level—estimating, for example, the effect of individual shootings, rather than the average effect using the synthetic control method (Abadie, Diamond, and Hainmueller Reference Abadie, Diamond and Hainmueller2015; Arkhangelsky et al. Reference Arkhangelsky, Athey, Hirshberg, Imbens and Wager2021; Kreif et al. Reference Kreif, Grieve, Hangartner, Turner, Nikolova and Sutton2016; Porreca Reference Porreca2022). Similarly, if interested in whether a subset of observations drives results, Broderick, Giordano, and Meager (Reference Broderick, Giordano and Meager2020) have an implementable procedure.Footnote 47

Third, in some applications, implementing different modeling strategies for estimating treatment effects with longitudinal data may make sense. For example, others have attempted to bracket treatment effects of interest in longitudinal data. Specifically, Hasegawa, Webster, and Small (Reference Hasegawa, Webster and Small2019) recommend bracketing by splitting the control group into different groups based on relative comparisons to the treatment group. As they describe, “the basic idea is to consider one control group that has a lower expected outcome than the treated group in the before period and another control group that has a higher expected outcome than the treated group in the before period” (Hasegawa, Webster, and Small Reference Hasegawa, Webster and Small2019, 372). They show estimators using the lower control group and the higher control group bracket the causal effect of treatment. This approach may be fruitful in cases where the pool of control units is large and easily split.Footnote 48 Likewise, Ding and Li (Reference Ding and Li2019) and Angrist and Pischke (Reference Angrist and Pischke2008) suggest bracketing strategies combining the standard difference-in-differences specification with lagged dependent variables.Footnote 49 , Footnote 50

Fourth, scholars may be interested in estimating distributional effects. Recent work combines difference-in-differences estimators with those quantile regression produces (see Callaway and Li Reference Callaway and Li2019; Roth et al. Reference Roth, Sant’Anna, Bilinski and Poe2022). Using this test, HHB show little signs of shifts at any point along the distribution of Democratic vote shares—that is, there is little evidence of polarizing effects in Democratic and Republican counties (see their Supplementary Figure A12 and surrounding discussion).

Finally, in making modeling decisions, one should acknowledge the tradeoffs between bias and precision and the importance of considering power in testing for pre-trends (Freyaldenhoven, Hansen, and Shapiro Reference Freyaldenhoven, Hansen and Shapiro2019; Roth Reference Roth2022; Roth et al. Reference Roth, Sant’Anna, Bilinski and Poe2022). For example, higher-order polynomials for unit-specific time trends require more power and may inflate standard errors. In our applied example, we have taken great care to pay attention to effect sizes, statistical significance, and the range of potential effects.

Overall, we note how to implement difference-in-differences designs depends, to a certain extent, on the nature of the data—that is, whether or not there are likely violations of parallel trends or unaccounted treatment effect heterogeneity or both. In the case of gun violence on electoral outcomes, the former appears to be key, while the latter less so. However, this may not always be the case. We think it is best to follow the suggestions outlined above to ensure inferences are not misleading. In our applied example, doing so reconciles why different studies using the same data arrived at vastly different conclusions regarding gun violence’s effects on electoral vote shares.

SYNTHESIZING EVIDENCE FROM MULTIPLE SPECIFICATIONS

We note here a few important points required to come to a conclusion about mass shootings’ effects on elections. In our particular example, there are many plausible models, a small number of which show statistically significant effects. Given the potential for bias and the role of researcher decisions, it is important for researchers to (1) address potential threats to inference outlined above, (2) be transparent about the effect of simple changes to model specification, and (3) take a “preponderance of evidence” rather than a “singular model” approach.Footnote 51

What does this mean in the mass shootings context? Though on occasion we see intermittent statistically significant effects, these effects are (1) much smaller than previous research (i.e., GMAL and Yousaf) claims and (2) not robust to reasonable alternative specifications under the control of researchers. Taking a “singular model” approach makes researchers vulnerable to mistaken inferences given researcher degrees of freedom in choosing a model specification. However, a “preponderance of evidence” approach provides considerable reasons to doubt mass shootings have any significant, systematic, or large effects.Footnote 52 Models accounting for parallel-trends violations, while not completely ruling out some potential smaller effects (although almost all of these effects are not statistically significant), provide little to no evidence that mass shootings cause large electoral change in the United States and instead provide compelling evidence consistent with null effects.

This can be seen by synthesizing four pieces of evidence. First, though a very small number of corrected models show much smaller positive effect on Democratic vote share, almost all are not statistically significant. Second, negative effects also show up regularly across the small, but reasonable, changes to model specification within researcher control. Figure 11 plots the distribution of effects and p-values for all difference-in-differences models and event-study models we estimated above. As shown in panel a, all coefficients from difference-in-differences models with trends are much smaller than GMAL’s TWFE. Some are positive and some are negative and the distribution spikes near zero. The average effect is 0.9 percentage points. Panel c shows that event-study models also spike at zero, with similar numbers of positive and negative effects. The average effect for all posttreatment periods is 0.4 percentage points and the average in the year immediately following treatment (i.e., period 0) is 0.07 percentage points. Third, when significant and positive effects do show up, these effects are not robust to slight variations in model specification within researchers’ reasonable control. Fourth, sensitivity analyses embracing uncertainty around exact parallel-trends departures show results are highly sensitive to even minimal reasonable departures. Hence, the preponderance of evidence suggests large effects are implausible and that modest positive (or negative) effects are anything but sure.

Figure 11. Distribution of All Effect Estimates and P-Values for Models with County Trends

Note: Distribution of all model estimates with trends in Figure 5 in the first row and then for all the event-study estimates in the article on the bottom row. The event-study coefficients are shown for periods 0–4 post-treatment. The left panel in each row shows coefficients (in percentage point units). The right panel in each row shows the distribution of p-values across model specifications. Takeaway: Once we account for potential violations of parallel trends, the effects of shootings spike around zero, are only rarely significant, are not robust to slight changes in model specification, and are sometimes positive and sometimes negative.

CONCLUSION

In reconciling research on the effects of mass shootings on electoral outcomes, our work also highlights the considerations we argue should become standard practice given the hazards of navigating difference-in-differences designs. In addition to resolving an important question, we hope our article sparks a more nuanced approach to estimating difference-in-differences models. If appropriately used, the checks outlined above will help researchers make better inferences using this common identification strategy.

The methodological contribution we provide applies best to cases where the treatment may not be fully exogenous or may vary in timing across units, and with larger sample sizes having more cross sections than time points. Instances departing from these may leverage approaches similar to those outlined above, but with unique features. Moreover, we have not explored some valuable aspects of panel data estimation recently developed for scenarios with very few treatment units (e.g., synthetic controls) that are valuable. Finally, our work is applied to a context where there is not currently, nor any prospect of a future, experimental baseline. While the econometric literature has long highlighted the value of the checks we run—through proofs, simulations, and other validation techniques—there is not yet (to the best of our knowledge) a comparison of difference-in-differences tools in our arsenal to a randomized baseline. Future work would do well to find other contexts where randomization is possible and add this benchmarking task to our suite of studies on this widely used method, as has been done with other methodological techniques (e.g., Arceneaux, Gerber, and Green Reference Arceneaux, Gerber and Green2006; Green et al. Reference Green, Leong, Kern, Gerber and Larimer2009).

Returning to the context of this study, America’s legacy of gun violence is heartbreaking and the thousands of deaths that occur from guns each year constitute a policy failure of epic proportions. Yet scholars have disagreed whether policymakers relative inaction occurs in spite of (or as a result of a lack of) an electoral response. While agreeing mass shootings do not effect voter turnout, scholars have come to vastly different conclusions about mass shootings’ effects on vote shares. We show that we cannot definitively conclude gun violence has any impact on Democratic vote shares (either positive or negative) and that previous work showing such relationships fails to navigate many of the pitfalls of difference-in-differences designs, specifically a failure of the parallel trends assumption. Moreover, even the most generous interpretation—i.e., entirely ignoring statistical uncertainty (something that prior work suggests should not be done; see, e.g., Stock and Watson Reference Stock and Watson2020)—suggests shootings have, at best, modest effects on Democratic vote share much smaller than prior research emphasizes. Looking across all robustness checks, we cannot conclude mass shootings of any kinds substantially affect election outcomes. Such a conclusion comes only from selecting highly sensitive results that are not robust.

Furthermore, these estimates are all local to the county in which the shooting occurred. Though all mass shootings are repugnant, they are (thankfully) relatively rare. Given shootings only occurred in 0.4% of counties (116 total; 11.6 per election) in HHB’s data, 0.4% of counties (115 total; 11.5 per election) in GMAL’s data, and only 0.5% of counties (72 total; 14.4 per election) in Yousaf’s data further emphasizes the limited impact shootings have on elections.Footnote 53 Ultimately, both the modest effect sizes in percentage points and their limited scope show mass shootings have little substantive consequence for election outcomes. Even if we take point estimates above at face value and ignore statistical uncertainty (something we certainly should not do), given the county-specific effect of the size observed and their infrequency mass shootings would have virtually no effect on any statewide or national election.

Lastly, as a reminder, there are often multiple ways of conceptualizing treatment exposure. These include, but are not limited to, short-lived treatments (constrained only to the period when they happen) or longer-term treatments (turned on in all periods after treatment occurs), spillover treatments (affecting units adjacent to treatment), varying dosage treatments, and even the possibility of national treatments drowning out any potential local effects. However, the best evidence we have currently suggests that none of these occur in mass shootings contexts (see HHB, 1377).

Our work sets the table for future work on the political economy of gun violence and retrospective voting/accountability more generally. Future work would do well to explore why mass shootings fail to substantively change the electoral incentives elected officials face, despite having favorable conditions to do so (HHB). Our work also acts a guide for researchers navigating the potential pitfalls of difference-in-difference designs. Future work would do well to continue to advance the boundaries of this promising and increasingly common method for making causal inferences.

SUPPLEMENTARY MATERIAL

To view supplementary material for this article, please visit https://doi.org/10.1017/S0003055424000108.

DATA AVAILABILITY STATEMENT

Research documentation and data that support the findings of this study are openly available at the American Political Science Review Dataverse: https://doi.org/10.7910/DVN/GH69TI.

Acknowledgments

We would like to thank Laura García-Montoya, Ana Arjona, Matthew Lacombe, and Hasin Yousaf for making their replication data and code available. We also thank Matt Baldwin, Charles Crabtree, Bernard Fraga, Andrew Goodman-Bacon, Justin Grimmer, Zoltan Hajnal, Nazita Lajevardi, Gareth Nellis (and the UCSD Graduate Student Seminar), Ben Newman, Tyler Reny, Brian Shaffner, Zayne Sember, Neil Visalvanich, Bertrand Wilden, and Yiqing Xu, and five anonymous reviewers for their feedback.

CONFLICT OF INTEREST

The authors declare no ethical issues or conflicts of interest in this research.

ETHICAL STANDARDS

The authors affirm this research did not involve human participants.

Footnotes

1 Likely because mass shootings have disproportionately occurred in recent years and in growing population areas (Musu-Gillette et al. Reference Musu-Gillette, Zhang, Wang, Kemp, Diliberti and Oudekerk2018; U.S. Government Accountability Office 2020) coinciding with political realignments where these areas have become more Democratic (DeSilver Reference DeSilver2016).

2 While stating “[their] findings hold when [they] replicate [Hassell, Holbein, and Baldwin Reference Hassell, Holbein and Baldwin2020; hereafter HHB] models with [their] data” (17) which include county-specific time trends, García-Montoya, Arjona, and Lacombe Reference García-Montoya, Arjona and Lacombe2022 (GMAL) do not provide models with county-specific time trends in any of their models in their manuscript, appendix, or replication materials. Yousaf (Reference Yousaf2021) also does not include models with time trends. None, including HHB, apply more recent advances addressing potential parallel-trends violations.

3 Roth et al. (Reference Roth, Sant’Anna, Bilinski and Poe2022, 2219) also note a related assumption requiring “the treatment has no causal effect before its implementation (no anticipation).”

4 Tests of parallel trends are a part of a broader form of falsification testing (e.g., Keele Reference Keele2015; Keele and Minozzi Reference Keele and Minozzi2013), wherein one provides evidence against the validity of an identification strategy (acknowledging that evidence for that identification strategy is not possible). Even without appearing to fail parallel trends assumptions, issues can arise, for example, with compound treatments (Keele and Minozzi Reference Keele and Minozzi2013). Moreover, parallel-trends tests can be underpowered (Bilinski and Hatfield Reference Bilinski and Hatfield2018; Rambachan and Roth Reference Rambachan and Roth2021).

5 HHB also use regression discontinuity in time to assess shootings’ effects on voter registration (also a null result).

6 Depending on the sample one uses—either the full pool of observations or only ones where covariates are available—GMAL’s naive TWFEs are 5.5 and 4.5 percentage points, respectively ( $ p<0.001 $ for both). GMAL run various specifications, which range in terms of the size of effects they document; being somewhere between 2 and 5 percentage points (see their Figure 4).

7 Unlike HHB, neither GMAL nor Yousaf examine midterms or state and local races.

8 HHB focus on school shootings occurring between 2006 and 2014 and, in robustness checks, between 2000 and 2018, GMAL focus on “rampage-style” shootings between 1980 and 2016, and Yousaf uses the FBI’s definition of a mass shooting “leading to four or more deaths at one location” with data from 2000 to 2016.

9 We are grateful to all the authors because we successfully replicated all reported findings using their code. As discussed below, our finding that mass shooting effects are insubstantial when one fully accounts for parallel trends is also corroborated by Marsh (Reference Marsh2022). While Marsh does provide some evidence that mass shootings close to an election have a slight positive effect on turnout (see Marsh Reference Marsh2022, Figure 1), HHB show that the effects of school shootings close to an election on turnout are highly sensitive to model specification (see HHB, Figure A7), a pattern also somewhat evident in Marsh’s models (see Marsh Reference Marsh2022, Table E2 and E5). Importantly, then, given the lack of any substantive effect on turnout, any increase in Democratic vote share should come from persuasion, rather than mobilization, unless gun violence simultaneously demobilizes Republicans and mobilize Democrats at the exact same rates, which is highly unlikely. However, any persuasive effect would also likely show up in attitudinal shifts and previous research on the attitudinal effects of mass shootings has disagreed whether attitudinal effects are present and, if they are, whether these effects are polarizing or a uniform leftward shift (Barney and Schaffner Reference Barney and Schaffner2019; Hartman and Newman Reference Hartman and Newman2019; Rogowski and Tucker Reference Rogowski and Tucker2019). An absence of an attitudinal shift does not alone undermine GMAL and Yousaf’s results (attitudes are not behaviors) but it provides a theoretical reason to question effects on vote shares.

10 These papers also include some time-varying controls, but the bulk of identifying assumptions come from county and year fixed effects. In some specifications, Yousaf compares successful shootings with non-successful shootings and in others includes flexible population time trends. GMAL, in some specifications, use neighboring counties as the control group, state fixed effects instead of county fixed effects, or decade fixed effects as opposed to year fixed effects.

11 HHB include unit-specific time trends.

12 Another researcher decision (of less consequence here) is how to estimate standard errors (MacKinnon, Nielsen, and Webb Reference MacKinnon, Nielsen and Webb2023). We cluster standard errors at the treatment level (the county level). In general, we advise thoughtfulness in clustering standard errors (see Abadie et al. Reference Abadie, Athey, Imbens and Wooldridge2023; Cameron and Miller Reference Cameron and Miller2015).

13 As HHB note, treatment does not need to be constrained to counties where shootings occur. HHB examine (and fail to find) effects in surrounding counties, as a function of the distance to a shooting, as a function of the severity of a shooting, and at the national level (with daily voter registration counts as the outcome).

14 Supplementary Figures S13 and S14 provide visual illustrations of these two approaches (for a random sample of the observations) using the panelView package developed by Mou, Liu, and Yiqing (Reference Mou, Liu and Yiqing2022a).

15 Table R1 in the Dataverse Files for this project provides estimates for Figure 1 (see Hassell and Holbein Reference Holbein and Hassell2024).

16 Marcus and Sant’Anna (Reference Marcus and Sant’Anna2021) note whether pretreatment tests validate parallel trends assumptions “depends on the chosen [parallel] trends assumption.” For other discussions of pre-trends tests, see Kahn-Lang and Lang (Reference Kahn-Lang and Lang2020), Bilinski and Hatfield (Reference Bilinski and Hatfield2018), and Roth et al. (Reference Roth, Sant’Anna, Bilinski and Poe2022), and for potential relaxations of parallel trends assumptions, see Manski and Pepper (Reference Manski and Pepper2018), Rambachan and Roth (Reference Rambachan and Roth2021), and Freyaldenhoven, Hansen, and Shapiro (Reference Freyaldenhoven, Hansen and Shapiro2019).

17 GMAL do this but fail to recognize potential violations of parallel trends in the bottom four figures for shootings after 2000 in GMAL’s Supplementary Section A.4. Visual inspection is important but seldom sufficient.

18 Figure 2 uses GMAL’s data. Supplementary Figure S10 shows Yousaf and HHB data.

19 For the few counties with multiple shootings, we code the first shooting and treat all subsequent years as posttreatment.

20 See, e.g., GMAL’s Supplementary Section A.4 compared to Figure 2 presented here.

21 Tests for imbalance are context-specific and scholars should consider the necessary precision of an imbalance using equivalence testing (Hartman and Hidalgo Reference Hartman and Hidalgo2018).

22 Supplementary Tables S13 and S14 provide the table version of panels a and b.

23 HHB control for pretreatment trends, but TWFE estimate with their data also produces 2.2 percentage point increases in Democratic vote share 4 years prior ( $ \beta =0.022 $ , $ p<0.073 $ ).

24 Yousaf’s time frame only permits examining three Presidential election cycles prior.

25 In another approach, we change the dependent variable to the change in Democratic vote share from the election before shootings occurred to the election in which counties are actually treated thus skirting Nickell bias arising from models with lags and fixed effects (Beck, Katz, and Mignozzetti Reference Beck, Katz and Mignozzetti2014).

26 Supplementary Tables S19–S22 provide model estimates for these figures and Supplementary Figure S14 shows results when coding counties with shootings as not reverting to the control afterward.

27 Supplementary Table S11 shows including covariates with a linear time trends makes effects even smaller—0.2 percentage points ( $ p=0.596 $ ; 95% CI: [ $ -0.7 $ , 1.2]).

28 Using the second treatment coding, the effect is 7.3 times smaller.

29 As shown in Supplementary Figure S14, Yousaf’s effects are not significant in three-fourths of the models run with the second treatment coding.

30 Adding time trends might artificially inflate standard errors to unpalatable levels. However, confidence intervals remain small in models with linear time trends. They are slightly less precise with quadratic time trends, but are, to our eye, still quite tight with higher-order specifications shown in the Supplementary Material.

31 Moreover, the null effects (once accounting for time trends) continue in robustness checks run by GMAL and Yousaf. Adding linear or quadratic time trends to Yousaf’s original comparison of shootings versus failed shootings, reduces estimates ranging from 0.04 to 1.4 percentage points, with none close to statistical significance. Effects are also smaller and insignificant after adding linear or quadratic time trends to GMAL’s original estimates using neighboring counties ( $ \beta $ = 0.5 percentage points; $ p<0.61 $ and $ \beta $ = 0.5 percentage points; $ p<0.6 $ , respectively). GMAL also run model with state and decade fixed effects (which do not absorb county-specific factors). Effects also attenuate dramatically if we add trends to these models.

32 Supplementary Tables S23–S26 provide coefficients for these figures.

33 This approach codes treatment only in the time period when the shooting occurred.

34 In the GMAL data, the evidence for effects shows up in “rampage-style” shootings with killings ( $ \beta =1.0 $ percentage points; $ p=0.064 $ ; 95% CI: [−0.05, 2.1]) rather than “rampage-style” shootings without killings ( $ \beta =0.45 $ percentage points; $ p=0.550 $ ; 95% CI: [−1.0, 1.9]).

35 Quadratic county-specific trends are: $ -0.09 $ percentage points ( $ p=0.927 $ ; 95% CI: [−2.1, 1.90]).

36 This approach treats all units posttreatment as treated.

37 See also Supplementary Tables S27–S30.

38 This approach codes treatment only in the proximate time period.

39 Under this approach, “the parameter M governs the amount by which the slope … can change between consecutive periods, and thus bounds the discrete analog of the second derivative” (Rambachan and Roth Reference Rambachan and Roth2021, 13).

40 In addition, researchers can see how robust effects are to (unobserved) posttreatment departures of parallel trends by benchmarking to (observed) maximum pretreatment violations of parallel trends. This is the $ RM $ or “relative magnitudes” approach. Researchers choose different values of $ \overline{M} $ , which measures how much of the maximum pretreatment violation of parallel trends would lead effects to include null effects in the confidence set. Rambachan and Roth argue this approach is reasonable—for example, if “possible violations of parallel-trends are driven by confounding … (shocks) of a similar magnitude to confounding … shocks in the pre-period” (Rambachan and Roth Reference Rambachan and Roth2021, 12). Further, researchers can combine approaches in the $ SDRM $ condition. This approach “assume[s] … possible non-linearities in the post-treatment difference in trends are bounded by the observed non-linearities in the pre-treatment difference in trends” (Rambachan and Roth Reference Rambachan and Roth2021, 13).

Under this approach, $ \overline{M} $ is the parameter the research varies, allowing researchers to set “bounds [for] the maximum deviation from a linear trend in the post-treatment period by $ \overline{M}\ge 0 $ times the equivalent maximum in the pre-treatment period.” $ SDRM $ is similar to $ SD $ , but “allows the magnitude of … possible non-linearity to explicitly depend on … observed pre-trends” (Rambachan and Roth Reference Rambachan and Roth2021, 13).

41 The rationale is explained in Goodman-Bacon (Reference Goodman-Bacon2021). See also Cunningham (Reference Cunningham2021, chap. 9).

42 Here, we balance the panel and code posttreatment units as treated.

43 Weights for Goodman–Bacon decomposition are in Supplementary Tables S33 and S34. Supplementary Tables S8 and S9 show results using de Chaisemartin, D’Haultfoeuille, and Guyonvarch’s (Reference de Chaisemartin, D’Haultfoeuille and Guyonvarch2019) alternative approach.

44 With this approach, we make the panel strongly balanced and use the treatment in the current period coding.

45 Results are robust to excluding never-treated observations.

46 Only HHB consider these different types of coding treatment. They find null effects in all once accounting for time trends.

47 See their zaminfluenceR package.

48 Implementing bracketing in our case, the bracket effects between 0.37 and 1.0 percentage points in the model with quadratic unit trends—still much smaller than prior estimates.

49 As Ding and Li (Reference Ding and Li2019, 605) explain “for a true positive effect, if ignorability is correct, then mistakenly assuming parallel-trends will overestimate the effect; in contrast, if the parallel trends assumption is correct, then mistakenly assuming ignorability will underestimate the effect.”

50 For our case, the bracket is between 1 percentage point and 2.7 percentage points—much smaller than prior estimates. Moreover, this estimate is from the lagged dependent variable model without fixed effects. If we include the lagged dependent variable with fixed effects, the bounds are between 1 and 1.08 percentage points.

51 In comparing the various approaches, leveraging new programming tools that make estimating multiple approaches at once easier may be useful, such as Hollenbach (Reference Hollenbach2021).

52 Arguing scholars should run many model specifications may prompt issues with multiple comparisons. Scholars should be careful not to overinterpret isolated significant effects in a deluge of otherwise insignificant results. However, properly adjusting for multiple comparisons across similar robustness checks is not well developed. Moreover, in arguing for the null, it is more conservative not to make any adjustments for multiple comparisons.

53 Only HHB consider whether there are spillover effects on adjacent counties, and find none (see 1377).

References

REFERENCES

Abadie, Alberto, Athey, Susan, Imbens, Guido W., and Wooldridge, Jeffrey M.. 2023. “When Should You Adjust Standard Errors for Clustering?Quarterly Journal of Economics 138 (1): 135.CrossRefGoogle Scholar
Abadie, Alberto, Diamond, Alexis, and Hainmueller, Jens. 2015. “Comparative Politics and the Synthetic Control Method.” American Journal of Political Science 59 (2): 495510.CrossRefGoogle Scholar
Angrist, Joshua D., and Pischke, Jörn-Steffen. 2008. Mostly Harmless Econometrics. Princeton, NJ: Princeton University Press.CrossRefGoogle Scholar
Angrist, Joshua D., and Pischke, Jörn-Steffen. 2010. “The Credibility Revolution in Empirical Economics: How Better Research Design Is Taking the Con Out of Econometrics.” Journal of Economic Perspectives 24 (2): 330.CrossRefGoogle Scholar
Arceneaux, Kevin, Gerber, Alan S., and Green, Donald P.. 2006. “Comparing Experimental and Matching Methods Using a Large-Scale Voter Mobilization Experiment.” Political Analysis 14 (1): 3762.CrossRefGoogle Scholar
Arkhangelsky, Dmitry, Athey, Susan, Hirshberg, David A., Imbens, Guido W., and Wager, Stefan. 2021. “Synthetic Difference-in-Differences.” American Economic Review 111 (12): 4088–118.CrossRefGoogle Scholar
Armitage, Seth. 1995. “Event Study Methods and Evidence on Their Performance.” Journal of Economic Surveys 9 (1): 2552.CrossRefGoogle Scholar
Bai, Jushan, and Ng, Serena. 2021. “Matrix Completion, Counterfactuals, and Factor Analysis of Missing Data.” Journal of the American Statistical Association 116 (536): 1746–63.CrossRefGoogle Scholar
Baker, Andrew C., Larcker, David F., and Wang, Charles C. Y.. 2022. “How Much Should We Trust Staggered Difference-in-Differences Estimates?.” Journal of Financial Economics 144 (2): 370–95.CrossRefGoogle Scholar
Barney, David J., and Schaffner, Brian F.. 2019. “Reexamining the Effect of Mass Shootings on Public Support for Gun Control.” British Journal of Political Science 49 (4): 1555–65.CrossRefGoogle Scholar
Beck, Nathaniel L., Katz, Jonathan N., and Mignozzetti, Umberto G.. 2014. “Of Nickell Bias and Its Cures: Comment on Gaibulloev, Sandler, and Sul.” Political Analysis 22 (2): 274–8.CrossRefGoogle Scholar
Berge, Laurent, Krantz, Sebastian, and McDermott, Grant. 2022. “Fixest: Fast Fixed-Effects Estimations.” CRAN.Google Scholar
Bilinski, Alyssa, and Hatfield, Laura A.. 2018. “Nothing to See Here? Non-Inferiority Approaches to Parallel Trends and Other Model Assumptions.” Preprint, arXiv:1805.03273.Google Scholar
Binder, John. 1998. “The Event Study Methodology since 1969.” Review of Quantitative Finance and Accounting 11 (2): 111–37.CrossRefGoogle Scholar
Borusyak, Kirill, Jaravel, Xavier, and Spiess, Jann. 2021. “Revisiting Event Study Designs: Robust and Efficient Estimation.” Preprint, arXiv:2108.12419.Google Scholar
Broderick, Tamara, Giordano, Ryan, and Meager, Rachael. 2020. “An Automatic Finite-Sample Robustness Metric: Can Dropping a Little Data Change Conclusions?” Preprint, arXiv:2011.14999.Google Scholar
Callaway, Brantly, and Li, Tong. 2019. “Quantile Treatment Effects in Difference in Differences Models with Panel Data.” Quantitative Economics 10 (4): 1579–618.CrossRefGoogle Scholar
Callaway, Brantly, and Sant’Anna, Pedro H. C.. 2021. “Difference-in-Differences with Multiple Time Periods.” Journal of Econometrics 225 (2): 200–30.CrossRefGoogle Scholar
Cameron, A. Colin, and Miller, Douglas L.. 2015. “A Practitioner’s Guide to Cluster-Robust Inference.” Journal of Human Resources 50 (2): 317–72.CrossRefGoogle Scholar
Clarke, Damian, and Tapia-Schythe, Kathya. 2021. “Implementing the Panel Event Study.” Stata Journal 21 (4): 853–84.CrossRefGoogle Scholar
Cunningham, Scott. 2021. Causal Inference: The Mixtape. New Haven, CT: Yale University Press.Google Scholar
De Chaisemartin, Clément, and d’Haultfoeuille, Xavier. 2020. “Two-Way Fixed Effects Estimators with Heterogeneous Treatment Effects.” American Economic Review 110 (9): 2964–96.CrossRefGoogle Scholar
De Chaisemartin, Clément, and D’Haultfoeuille, Xavier. 2023. “Two-Way Fixed Effects and Differences-in-Differences with Heterogeneous Treatment Effects: A Survey.” The Econometrics Journal 26 (3): C1–C20.CrossRefGoogle Scholar
de Chaisemartin, Clément, D’Haultfoeuille, Xavier, and Guyonvarch, Yannick. 2019. “DID_MULTIPLEGT: Stata Module to Estimate Sharp Difference-in-Difference Designs with Multiple Groups and Periods.” Statistical Software Components S458643, Boston College Department of Economics.Google Scholar
DeSilver, Drew. 2016. “The Growing Democratic Domination of Nation’s Largest Counties.” Pew Research Center, July 21.Google Scholar
Ding, Peng, and Li, Fan. 2019. “A Bracketing Relationship between Difference-in-Differences and Lagged-Dependent-Variable Adjustment.” Political Analysis 27 (4): 605–15.CrossRefGoogle Scholar
Freyaldenhoven, Simon, Hansen, Christian, Pérez, Jorge Pérez, and Shapiro, Jesse. 2022. “Xtevent: Stata Module to Estimate and Visualize Linear Panel Event-Study Models.” Working Paper.Google Scholar
Freyaldenhoven, Simon, Hansen, Christian, Pérez, Jorge Pérez, and Shapiro, Jesse M.. 2021. “Visualization, Identification, and Estimation in the Linear Panel Event-Study Design.” Working Paper, National Bureau of Economic Research.CrossRefGoogle Scholar
Freyaldenhoven, Simon, Hansen, Christian, and Shapiro, Jesse M.. 2019. “Pre-Event Trends in the Panel Event-Study Design.” American Economic Review 109 (9): 3307–38.CrossRefGoogle Scholar
García-Montoya, Laura, Arjona, Ana, and Lacombe, Matthew. 2022. “Violence and Voting in the United States: How School Shootings Affect Elections.” American Political Science Review 116 (3): 807–26.CrossRefGoogle Scholar
Goodman-Bacon, Andrew. 2019. “So You’ve Been Told to Do My Difference-in-Differences Thing: A Guide.” Working Paper, Vanderbilt University.Google Scholar
Goodman-Bacon, Andrew. 2021. “Difference-in-Differences with Variation in Treatment Timing.” Journal of Econometrics 225 (2): 254–77.CrossRefGoogle Scholar
Goodman-Bacon, Andrew, Goldring, Thomas, and Nichols, Austin. 2019. “BACONDECOMP: Stata Module to Perform a Bacon Decomposition of Difference-in-Differences Estimation.” Statistical Software Components S458676, Boston College Department of Economics.Google Scholar
Goss, Kristin. 2010. Disarmed: The Missing Movement for Gun Control in America. Princeton, NJ: Princeton University Press.Google Scholar
Green, Donald P., Leong, Terence Y., Kern, Holger L., Gerber, Alan S., and Larimer, Christopher W.. 2009. “Testing the Accuracy of Regression Discontinuity Analysis Using Experimental Benchmarks.” Political Analysis 17 (4): 400–17.CrossRefGoogle Scholar
Grimmer, Justin, Hersh, Eitan, Meredith, Marc, Mummolo, Jonathan, and Nall, Clayton. 2018. “Obstacles to Estimating Voter ID Laws’ Effect on Turnout.” Journal of Politics 80 (3): 1045–51.CrossRefGoogle Scholar
Hansen, Ben B., and Bowers, Jake. 2008. “Covariate Balance in Simple, Stratified and Clustered Comparative Studies.” Statistical Science 23(2): 219–36.CrossRefGoogle Scholar
Hartman, Erin, and Hidalgo, F. Daniel. 2018. “An Equivalence Approach to Balance and Placebo Tests.” American Journal of Political Science 62 (4): 1000–13.CrossRefGoogle Scholar
Hartman, Todd K., and Newman, Benjamin J.. 2019. “Accounting for Pre-Treatment Exposure in Panel Data: Re-Estimating the Effect of Mass Public Shootings.” British Journal of Political Science 49 (4): 1567–76.CrossRefGoogle Scholar
Hasegawa, Raiden B., Webster, Daniel W., and Small, Dylan S.. 2019. “Evaluating Missouri’s Handgun Purchaser Law: A Bracketing Method for Addressing Concerns about History Interacting with Group.” Epidemiology 30 (3): 371–9.CrossRefGoogle Scholar
Hassell, Hans J. G., Holbein, John B., and Baldwin, Matthew. 2020. “Mobilize for Our Lives? School Shootings and Democratic Accountability in US Elections.” American Political Science Review 114 (4): 1375–85.CrossRefGoogle Scholar
Healy, Andrew, and Lenz, Gabriel S.. 2017. “Presidential Voting and the Local Economy: Evidence from Two Population-Based Data Sets.” Journal of Politics 79 (4): 1419–32.CrossRefGoogle Scholar
Holbein, John B., and Hassell, Hans J. G.. 2024. “Replication Data for: Navigating Potential Pitfalls in Difference-in-Differences Designs: Reconciling Conflicting Findings on Mass Shootings’ Effect on Electoral Outcomes.” Harvard Dataverse. Dataset. https://doi.org/10.7910/DVN/GH69TI.CrossRefGoogle Scholar
Hollenbach, Florian. 2021. “Comparing Staggered DiD.” https://github.com/fhollenbach/did_compare/blob/main/ComparingDiD.md.Google Scholar
Kahn-Lang, Ariella, and Lang, Kevin. 2020. “The Promise and Pitfalls of Differences-in-Differences: Reflections on 16 and Pregnant and Other Applications.” Journal of Business & Economic Statistics 38 (3): 613–20.CrossRefGoogle Scholar
Keele, Luke. 2015. “The Statistics of Causal Inference: A View from Political Methodology.” Political Analysis 23 (3): 313–35.CrossRefGoogle Scholar
Keele, Luke, and Minozzi, William. 2013. “How Much Is Minnesota Like Wisconsin? Assumptions and Counterfactuals in Causal Inference with Observational Data.” Political Analysis 21 (2): 193216.CrossRefGoogle Scholar
Kreif, Noémi, Grieve, Richard, Hangartner, Dominik, Turner, Alex James, Nikolova, Silviya, and Sutton, Matt. 2016. “Examination of the Synthetic Control Method for Evaluating Health Policies with Multiple Treated Units.” Health Economics 25 (12): 1514–28.CrossRefGoogle ScholarPubMed
Liu, Licheng, Wang, Ye, and Xu, Yiqing. 2024. “A Practical Guide to Counterfactual Estimators for Causal Inference with Time-Series Cross-Sectional Data.” American Journal of Political Science 68 (1): 160–76.CrossRefGoogle Scholar
Luca, Michael, Malhotra, Deepak, and Poliquin, Christopher. 2020. “The Impact of Mass Shootings on Gun Policy.” Journal of Public Economics 181: 104083.CrossRefGoogle Scholar
MacKinnon, James G., Nielsen, Morten Ørregaard, and Webb, Matthew D.. 2023. “Testing for the Appropriate Level of Clustering in Linear Regression Models.” Journal of Econometrics 235 (2): 2027–56.CrossRefGoogle Scholar
Manski, Charles F., and Pepper, John V.. 2018. “How Do Right-to-Carry Laws Affect Crime Rates? Coping with Ambiguity Using Bounded-Variation Assumptions.” Review of Economics and Statistics 100 (2): 232–44.CrossRefGoogle Scholar
Marcus, Michelle, and Sant’Anna, Pedro H. C.. 2021. “The Role of Parallel Trends in Event Study Settings: An Application to Environmental Economics.” Journal of the Association of Environmental and Resource Economists 8 (2): 235–75.CrossRefGoogle Scholar
Marsh, Wayde Z. C. 2022. “Trauma and Turnout: The Political Consequences of Traumatic Events.” American Political Science Review 117 (3): 1036–52.CrossRefGoogle Scholar
Mou, Hongyu, Liu, Licheng, and Yiqing, Xu. 2022a. “Package ‘panelView.’”Google Scholar
Mou, Hongyu, Liu, Licheng, and Yiqing, Xu. 2022b. “panelView: Panel Data Visualization in R and Stata.” SSRN 4202154.CrossRefGoogle Scholar
Musu-Gillette, Lauren, Zhang, Anlan, Wang, Ke, Kemp, Jana, Diliberti, Melissa, and Oudekerk, Barbara A.. 2018. “Indicators of School Crime and Safety: 2017.” Report, National Center for Educational Statistics (NCES 2018-036).Google Scholar
Olea, Montiel, Luis, José, and Plagborg-Møller, Mikkel. 2019. “Simultaneous Confidence Bands: Theory, Implementation, and an Application to SVARs.” Journal of Applied Econometrics 34 (1): 117.CrossRefGoogle Scholar
Porreca, Zachary. 2022. “Synthetic Difference-in-Differences Estimation with Staggered Treatment Timing.” Economics Letters 220: 110874.CrossRefGoogle Scholar
Rambachan, Ashesh, and Roth, Jonathan. 2021. “An Honest Approach to Parallel Trends.” Unpublished Manuscript, Harvard University.Google Scholar
Rambachan, Ashesh, Roth, Jonathan, and Bravo, Mauricio Caceres. 2021. “HonestDiD.” https://github.com/asheshrambachan/HonestDiD/tree/bc576d5f338dd01ebb641f30215792c0605bc08a.Google Scholar
Rogowski, Jon C., and Tucker, Patrick D.. 2019. “Critical Events and Attitude Change: Support for Gun Control after Mass Shootings.” Political Science Research and Methods 7 (4): 903–11.CrossRefGoogle Scholar
Rossin-Slater, Maya, Schnell, Molly, Schwandt, Hannes, Trejo, Sam, and Uniat, Lindsey. 2020. “Local Exposure to School Shootings and Youth Antidepressant Use.” Proceedings of the National Academy of Sciences 117 (38): 23484–9.CrossRefGoogle ScholarPubMed
Roth, Jonathan. 2022. “Pre-Test with Caution: Event-Study Estimates after Testing for Parallel Trends.” American Economic Review: Insights 4(3): 305–22.Google Scholar
Roth, Jonathan, Sant’Anna, Pedro H. C., Bilinski, Alyssa, and Poe, John. 2022. “What’s Trending in Difference-in-Differences? A Synthesis of the Recent Econometrics Literature.” Journal of Econometrics 235 (2): 2218–44.CrossRefGoogle Scholar
Schmidheiny, Kurt, and Siegloch, Sebastian. 2023. “On Event Studies and Distributed Lags in Two-Way Fixed Effects Models: Identification, Equivalence, and Generalization.” Journal of Applied Econometrics 38(5), 695713.CrossRefGoogle Scholar
Sides, John, Vavreck, Lynn, and Warshaw, Christopher. 2022. “The Effect of Television Advertising in United States Elections.” American Political Science Review 116 (2): 702–18.CrossRefGoogle Scholar
Stock, James H., and Watson, Mark W.. 2020. Introduction to Econometrics. New York: Pearson.Google Scholar
Sun, Liyang. 2021. “EVENTSTUDYINTERACT: Stata Module to Implement the Interaction Weighted Estimator for an Event Study.” Statistical Software Components S458978, Boston College Department of Economics.Google Scholar
Sun, Liyang, and Abraham, Sarah. 2021. “Estimating Dynamic Treatment Effects in Event Studies with Heterogeneous Treatment Effects.” Journal of Econometrics 225 (2): 175–99.CrossRefGoogle Scholar
U.S. Government Accountability Office. 2020. “K-12 Education: Characteristics of School Shootings.” GAO-20-455.Google Scholar
Wing, Coady, Simon, Kosali, and Bello-Gomez, Ricardo A.. 2018. “Designing Difference in Difference Studies: Best Practices for Public Health Policy Research.” Annual Review of Public Health 39: 453–69.CrossRefGoogle ScholarPubMed
Xu, Yiqing. 2023. “Causal Inference with Time-Series Cross-Sectional Data: A Reflection.” In Oxford Handbook of Engaged Methodological Pluralism in Political Science, Vol. 1, eds. Janet M. Box-Steffensmeier, Dino P. Christenson, and Valeria Sinclair-Chapman, online edition. https://doi.org/10.1093/oxfordhb/9780192868282.013.30.CrossRefGoogle Scholar
Yousaf, Hasin. 2021. “Sticking to One’s Guns: Mass Shootings and the Political Economy of Gun Control in the United States.” Journal of the European Economic Association 19 (5): 2765–802.CrossRefGoogle Scholar
Zhang, Shuo. 2022. “DIDmultiplegt: Estimation in DID with Multiple Groups and Periods.” CRAN.Google Scholar
Figure 0

Figure 1. Differences in Previous Studies’ Estimated Effect of Mass Shootings on Election Outcomes Are Not Driven by Data ChoicesNote: Estimates include county and year fixed effects (i.e., the TWFE estimator) with standard errors clustered at the county level and no covariates. The top panel shows effect estimates coding only the election immediately after a shooting occurs as having been treated; the bottom panel considers all post-shooting elections in counties with a shooting as treated. Coefficients, standard errors, and p-values are labeled for each coefficient. Takeaway: Naive TWFE estimators suggest mass shootings—regardless of the data/coding used—increase Democratic vote share in the county where shootings happens by 2.6–8.7 percentage points.

Figure 1

Figure 2. Trends in Presidential Vote in Counties with Mass Shootings Prior to Shootings, Compared to Trends in Counties without ShootingsNote: Pretreatment trends of Democratic vote share in counties where a shooting occurred (left panel) benchmarked to the trends in Democratic vote share found in counties where a shooting did not occur (right panel). Lighter lines show the patterns of individual counties; darker lines show the overall pattern for all counties. Takeaway: Counties that have shootings trended more Democratic even before the shootings occurred, whereas counties without a shooting trended slightly more Republican. Models that do not account for differential trends across counties will be biased.

Figure 2

Figure 3. The Effect of Shootings on Election Outcomes Many Years BeforeNote: Effect of mass shootings on Democratic vote share in the years prior to when a shooting occurred. Treatment #1 is coded such that only elections with shooting are coded as treated; Treatment #2 is coded such that all elections after a shooting occurs in a county are coded as treated. All models’ standard errors are clustered at the county level. Takeaway: TWFE estimators without time-trends indicate shootings may have an effect up to and including 20 years prior to when a shooting occurred.

Figure 3

Figure 4. Event-Study Estimates Show that TWFE Fails to Account for Pretreatment TrendsNote: Event-study estimates with county and year fixed effects (GMAL’s data). Baseline election year is shown with a gray dotted line. Following prior work, we bin our extreme points (Baker, Larcker, and Wang 2022; Schmidheiny and Siegloch 2019). Takeaway: Counties that have shootings trended more Democratic even before the shootings occurred. The increase that occurs after a shooting is entirely consistent with a general trend toward more Democratic election outcomes. Models that do not account for differential trends across counties will be biased.

Figure 4

Figure 5. Effects of Mass Shootings on Elections after Absorbing County-Specific TrendsNote: Effect of mass shootings of various types once we account for differential trends. Within each panel, the first three estimates are using the GMAL coding of mass shootings and their data, the next comes from HHB, and the last comes from Yousaf. For cubic and quartic specifications, see Supplementary Figure S9. For effects where we code all post-shooting counties as being treated—not just counties and years with shootings—see Supplementary Figure S14. Takeaway: Once we account for differential trends across counties, the effects of mass shootings are all smaller and precisely estimated.

Figure 5

Figure 6. Event-Study Estimates of Shootings after Absorbing County-Specific TrendsNote: Event-study estimates from the HHB and GMAL data with county and year fixed effects and county-specific quadratic time trends. These use the method developed by Freyaldenhoven et al. (2021) to account for pre-trends in event-study designs. Analysis executed using the xtevent and xteventplot commands in STATA (Freyaldenhoven et al. 2022). These commands, as a default, plot both the standard confidence intervals and those developed by Olea, Luis, and Plagborg-Møller (2019), which were developed for contexts with dynamic effects. The figure uses the same y-axis as Figure 4 for ease in comparing across the two. Takeaway: Once time trends are taken into account, the effect of shootings attenuates considerably.

Figure 6

Figure 7. Liu, Wang, and Xu (2024) Interactive Fixed Effects Counterfactual EstimatorNote: The interactive fixed effects counterfactual estimator developed by Liu, Wang, and Xu (2024) using GMAL’s data. Panel a shows the TWFE estimated by Liu, Wang, and Xu’s (2024) FECT package; it is analogous to Figure 5, but their procedure estimates slight differences—for example, in the number of pre- and posttreatment periods. In panel b, the number of factors (r) is set to 3—that chosen by cross-validation and the degree of the polynomial is set to 4. In the bottom row, r is set to 1 in both panels and degree 2 in panel c and 4 in panel d. For other variations, see the Supplementary Material. Takeaway: The upward trend in the TWFE model (i.e., panel a) is indicative of violation of the parallel trends assumption. In the interactive fixed effects models, there is no evidence of the substantial effects shown in more simplistic model specifications that do not account for potential violations of the parallel trends assumption.

Figure 7

Figure 8. Implementing Rambachan and Roth’s Sensitivity Analysis in the Shooting ExampleNote: Results from the sensitivity analysis suggested by Rambachan and Roth (2021) using GMAL’s data—that is, testing for effect sensitivity across $ {\triangle}^{SD}(M) $. The models incorporate information from three elections prior to treatment and five post-treatment periods. Takeaway: The results show the effects of shootings on vote shares are highly sensitive and do not hold with even minor deviations from parallel trends.

Figure 8

Figure 9. Illustration of the Goodman–Bacon Decomposition of the TWFE ModelsNote: This figure shows the results from the Bacon decomposition for the TWFE models. The figure also shows all of the possible $ 2\times 2 $ difference-in-differences (DiD) estimate, with their weights for the ATE on the x-axis and the effect size on the y-axis. The horizontal line shows the overall DiD estimate.

Figure 9

Figure 10. Sun and Abraham (2021) Approach for Estimating Dynamic Treatment Effects in Event Studies with Heterogeneous Treatment EffectsNote: Results from the clean comparisons suggested by Sun and Abraham (2021) using the GMAL and HHB data. Models include quadratic county-specific time trends to address potential violations of the parallel trends assumption in the TWFE. Takeaway: Clean comparison effects with trends show no sign of a sizable and durable effect on Democratic vote shares shown in the TWFE nor in the simple event-study plot (see Figure 5).

Figure 10

Figure 11. Distribution of All Effect Estimates and P-Values for Models with County TrendsNote: Distribution of all model estimates with trends in Figure 5 in the first row and then for all the event-study estimates in the article on the bottom row. The event-study coefficients are shown for periods 0–4 post-treatment. The left panel in each row shows coefficients (in percentage point units). The right panel in each row shows the distribution of p-values across model specifications. Takeaway: Once we account for potential violations of parallel trends, the effects of shootings spike around zero, are only rarely significant, are not robust to slight changes in model specification, and are sometimes positive and sometimes negative.

Supplementary material: File

Hassell and Holbein supplementary material

Hassell and Holbein supplementary material
Download Hassell and Holbein supplementary material(File)
File 2.1 MB
Submit a response

Comments

No Comments have been published for this article.