Have you ever wondered why some elections have higher voter turnout than others?
If you have, my new (open source) Political Behavior article with Ferran Martínez i Coma might be of interest.
How *and why* did we read over 300 turnout articles, extract 127 unique explanatory variables from them, collect data on 70 factors, and then run over 15 million regressions?
A very good question…one I asked myself a few times over the last few years.
This project began shortly after Ferran and I finished an article looking at how election dynamics shaped experts perceptions of electoral integrity.
For that article I played around with extreme bounds analysis (EBA) as a robustness check of our main models. We ran 164,220 additional regressions to see whether our results were sensitive to different model specifications. Nope, our main IVs remained robustly significant.
Now EBA is not a new method. Leamer (1985) proposed it as a form of organized sensitivity analysis to see whether a coefficient is robust to the inclusion of other explanatory variables.
A decade or so later, Sala-i-Martin (1997) suggested that Leamer’s extreme bounds threshold for robustness was overly strict and proposed a version of model averaging where we look at the entire distribution of betas instead of just the extreme values.
In the two decades since Sala-i-Martin (1997) and Levine and Renelt’s (1992) EBA studies of economic growth, the method has been used to evaluate the correlates of human rights violations (Hafner-Burton 2005), democracy (Gassebner et al. 2013), life expectancy and infant mortality (Carmignani et al. 2014), health care expenditures (Hartwig & Sturm 2014), human trafficking (Cho 2015), coups (Miller et al. 2018), amongst many others.
Therefore, we thought it would be interesting to use EBA again, this time at the heart of our research design. But what variables should we include in our models?
To make our list, @fmartinezicoma read over 300 turnout articles and chose 44 comparative turnout students that met our decision criteria.
These articles included 127 unique variables as either independent or control variables. We grouped these variables into three main types—institutional, political, and socioeconomic.
During this process, several trends emerged.
(1) Less than half (44%) of the 127 variables appear in more than one study.
(2) Among the 55 variables that appear in more than one article, over half (57%) were measured in more than one way.
(3) While voting is (of course) a profoundly political act, the most common empirically tested arguments in the literature are of institutional or socio-economic mechanisms.
It is clear that even with similar research questions (“what drives turnout?”) and the same International IDEA turnout data, researchers estimated wildly different models.
This is not unique to political science of course. The most dramatic example of this I’ve come across is Siberzahn et al.’s (2018) article “Many Analysts, One Dataset” where 29 teams of 61 analysts all use the same dataset to estimate soccer referees’ use of red cards.
Anyways, I set to work tracking down as many of these variables as possible. In the end, we used 70 national-level predictors of voter turnout. Including all 127 predictors discussed above in our empirical models was not possible due to data availability—several variables are available for only a short span of time, specific region (i.e., Europe) or a clear set of countries (i.e., OECD members).
We included five of the most common predictors in every model (lagged turnout, compulsory voting, proportional representation, per capita income, and population).
We then used different combinations of the other 65 variables in sets of three.
Our sample was 579 parliamentary elections in 80 countries, 1945-2014.
We (and the reviewers) realized that this first set of EBA models was not the end, but the beginning of our sensitivity research.
Indeed, Plümper and Traunmüller’s (2020) recent Monte Carlo study of these and other sensitivity tests demonstrates how there is no definition of “robustness” that completely avoids false negatives and false positives.
Therefore, we ran a series of models using fixed and random effects, different dependent variables, and different samples. All in all, we ran sixteen different sets of models.
Overall, we found that 22 variables are robustly associated with voter turnout, including compulsory voting, concurrent elections, competitive elections, inflation, previous turnout, and economic globalization.
I think our article provides some suggestive answers to our research question—what drives national-level voter turnout?
Of course, trying to answer one question leads to a host of others…
(1) Do different operationalizations of the same concept lead to different conclusions?
(2) Would theoretically informed interaction terms also make the grade?
(3) Plümper and Traunmüller (2018) find that Bayesian model averaging’s (BMA) definition minimizes false positives. Would a BMA of voter turnout lead to different results?
Anyways, I hope this article is of use to both (1) people interested in voter turnout and (2) all of us interested in making sure our results are not sensitive to different model specifications.
The paper is open-source and is online now although it is not likely to actually get in an issue for a few years. The replication files are also available now at my Dataverse.