Abstract
Billions of people outside of the US hold a vested interest in the outcome of US presidential elections, despite not being allowed to vote. We developed 9 Voting Advice Applications (VAAs) for the 2020 elections and 5 VAAs for the 2024 elections in partnership with leading newspapers across Europe, Asia, and Oceania. Across these two elections, the VAA tools offered 724,294 users from 187 countries an issue-based comparison with Biden and Trump in 2020, and with Trump and Harris in 2024. Users answered demographic questions, the issue statements, and additional attitudinal items. These issue statements covered both domestic topics and topics of the US’ foreign policy. The candidates’ positions were coded using robust primary and secondary sources on 29 (2020) and 62 (2024) issue statements. In this paper, we describe the candidates’ placement data as well as the user data generated by these VAAs, offering an unprecedented opportunity to analyze global perceptions of US politics and the election’s far-reaching influence on international relations.
Similar content being viewed by others
Background & Summary
The US presidential election is arguably the most salient foreign election worldwide. Beyond the US population, billions of people across the globe are impacted by the outcome of who wins the White House. The policies of the US president influence global security, trade, climate policy, and international cooperation, making the election’s implications far-reaching. Despite its importance, foreign perspectives on the US election remain understudied, even as scholars have come to grapple with the relevance of such perceptions in understanding the stability of US hegemony and the diffusion of ideology across borders. Collecting data on foreign opinions regarding the US election allows us to better understand how international audiences perceive American democracy, its candidates, and their policy positions.
To capture these perspectives, we developed Voting Advice Applications (VAAs) for the 2020 and 2024 US elections tailored to an international audience. These tools are available at usa2020.electioncompass.org and usa2024.electioncompass.org, respectively. VAAs are interactive tools that help users identify which candidate aligns most closely with their policy preferences. Traditionally used for national elections, VAAs also serve as valuable research instruments for gauging public opinion. Our VAAs present users with a series of Likert-scale issue statements reflecting key campaign issues, often with a particular international angle. The tool then compares their responses to the positions of Biden and Trump, and Trump and Harris respectively, calculating the proximity between the candidates and the user. The tools were developed by researchers at the Dutch Kieskompas – Election Compass research institute in collaboration with scholars at the University of Pennsylvania and the VU University Amsterdam. They were funded and disseminated by major news outlets across a range of countries as well as by the University of Amsterdam.
Our contribution is comprised of four main datasets. For both the 2020 and 2024 elections, we coded the respective candidates’ positions on 29 and 62 statements, respectively. Each candidate position includes a 5-point Strongly Agree to Strongly Disagree scale point, which is supported by up to three sources. In addition, again for both elections, we present the responses of 571.624 and 152.670 users, respectively. The 2020 tools were disseminated in Belgium, Germany, Denmark, France, the Netherlands, Romania, Bulgaria, and the US. While our ‘base’ VAA was tailored to an American audience, each disseminated version was customized with questions specifically relevant to the international audience it served. The 2024 tools were disseminated in Belgium, India, and by an Asia-wide newspaper, in addition to the tool tailored to the US. In total, the tools were accessed by users in 187 countries. The respective media partners and user totals are indicated in Table 1.
This dataset has the potential to inform policy debates and enhance academic understanding of international electoral perceptions and their downstream consequences. Specifically, it allows researchers to explore how foreign publics evaluate US candidates and their policy stances, shedding light on perceptions of US leadership and broader patterns of ideological diffusion across different regions and political contexts. Ultimately, these insights can be leveraged by policymakers, media organizations, and academics to refine international engagement strategies, assess the global reception of US foreign policy, and deepen our understanding of the international ramifications of US electoral outcomes.
Methods
Before detailing the data generation process, we briefly discuss the dual role of voting advice applications (VAAs) as both informational tools for users and instruments for collecting public opinion data. We then outline the development of the issue statements and the qualitative coding of the candidates’ positions. Finally, we describe the process by which the user data was generated through interaction with the VAAs.
Voting advice applications are online tools designed to provide voters with information regarding substantive overlap with parties or candidates running in a particular election. They emerged in the 1990s with the advent of the internet, and are now commonplace in many multiparty systems around the world1. In some countries, like the Netherlands, certain VAAs are consulted by up to half of the electorate2. However, these tools do not attract random subsets of the population, often leading to samples with an overrepresentation of young, highly educated males3,4 without an immigration background and a disproportionate level of political interest5. While such differential consultation may have attenuated over time6, it remains heavily conditioned by the level of internet penetration and usage in a country7. In addition, these tools are exposed to a large number of designer degrees of freedom. Both the wording of issue statements8,9 and the selection of included issues10 can affect the results that the tools provide. Subsequently, the particular algorithms that are used to match users with candidates based on their responses can introduce another layer of variability11,12. We should therefore understand the capacity of VAAs to provide voting advice in the context of their design decisions and the tools’ reach.
Aside from providing information to voters, these tools are unique data collection devices. The placements of the parties or candidates have been used in electoral analysis13, presenting insights into the estimated policy positions of these entities relative to each other14. The combination of public opinion data and party placements has been used in the study of electoral congruence15. Perhaps the largest advantage of VAAs over alternative data collection methods for public opinion data is their large reach. Like other nonprobability samples, VAA data faces selection bias risks, which occurs when opt-in occurs at differential rates16,17. Nevertheless, both the data and the panels that result from these VAAs have frequently been used in academic publications18,19,20. Given the appropriate statistical methods, VAA data can reliably be used for estimating voter policy positions21, providing evidence for the validity of VAA data as instruments for public opinion research.
The tools that collected the user data as presented here are designed by the Amsterdam-based Kieskompas - Election Compass research institute, which develops the second-most visited VAA during Dutch elections, and which has developed VAAs in over 30 countries. Previous work has outlined the approach that the research institute takes with regards to the development of their VAAs22,23. The following two sections provide a detailed account of the development of the VAAs for the 2020 and 2024 US elections.
Candidate positions
The first step in designing the candidate position data involves issue mapping and the drafting of issue statements. This process begins with the creation of a longlist of politically salient topics that are particularly relevant to foreign audiences. Coders then refine this list by merging overlapping issues, excluding those deemed non-salient, and ultimately producing a finalized shortlist of statements for inclusion in the VAA. For each selected issue, a corresponding statement is formulated. These statements are crafted to be as unambiguous as possible9, avoiding unnecessary quantification or qualification24, and are designed to capture ideological positions rather than specific policy proposals25. At this stage, each statement is also categorized as reflecting a culturally liberal, culturally conservative, economically left, or economically right position, aiming for a roughly equal number of each pole per dimension8. This careful formulation process ensures that the issue statements are both accessible to users and analytically useful for subsequent coding and analysis, producing 29 statements in 2020 and 62 statements in 2024.
Table 2 lists all statements used in the 2020 and 2024 VAAs across all newspapers. Table 3 lists all statements that the candidates were positioned on, but did not feature in all newspapers in their respective cycles. In the few cases where a statement appeared in only one publication, the name of that publication is indicated. Many statements were reused between 2020 and 2024, although at times with slight modifications as indicated. In 2020, 29 statements were coded in total, 26 of which were ultimately included in the VAA. The 2024 iteration drew from a larger set of coded statements (62). The VAAs distributed in Belgium and the United States included 25 statements, while the versions published in The Diplomat and The Hindu each had 30 statements. These additional statements included some of those developed during the initial issue formulation and placement coding process, and some developed specifically to address issues relevant to the audiences of these publications.
The two major-party candidates in both elections are positioned on all of the above statements. The placement coding process includes multiple steps to ensure that coders independently reach conclusions about how to position the candidates against each of the statements on the shortlist. In 2024, two coders independently completed the coding process, while a third coder resolved any discrepancies between the initial two coders. The intercoder reliability between the two initial coders was 0.78 (weighted Cohen’s kappa, ICC, and Krippendorf). In 2020, due to more limited funding, two coders evaluated only a subset of the statements. After reviewing the level of agreement and reliability within this subset, it was determined that a single coder could reliably code the remaining statements. During the placement coding process, coders find authoritative sources that clearly indicate the candidate’s position on a given issue. Sources are valued hierarchically, with official candidate platforms and candidate speeches and interviews preferred. We consider such sources to be primary sources. If these are not available for a given statement, coders then look for related positions or reliable media reports that either indicate the candidate’s position or allow us to infer their position with a high degree of confidence. Third, coders looked at candidate voting records. It should be noted that a significant proportion of Trump’s 2020 positions were sourced from Twitter (now X) given his heavy reliance on the platform while campaigning. All of these sources were weighted heavily based on their recency, with particular caution applied to evidence from previous administrations, which was considered less indicative of current positions. When sources provide conflicting information about a candidate’s position, coders evaluate them based primarily on recency and the hierarchy of sources. For each statement, we require a minimum of one source and a maximum of three to justify the coder’s positioning of the candidate. The statements feature five-point Likert scales and include a ‘No Opinion’ option. For a candidate to be coded as ‘Strongly Agree’ or ‘Strongly Disagree’ with a statement, we require evidence of unqualified support or opposition to the issue, respectively. Candidates can also be coded as ‘Agree’ or ‘Disagree’ in instances where they voice conditional agreement or disagreement. Candidates who offer highly nuanced positions are coded as ‘Neither Agree Nor Disagree,’ while candidates who have not voiced any position are coded as ‘No Opinion.’ Statements where either one of the candidates is coded as having ‘No Opinion’ will not be featured in the VAA, but we retain those statements in the candidate placement data we share.
To ensure clear differentiation between the two candidates in the VAA, issues were selected and worded to maximize the contrast in their coded positions. This resulted in a higher proportion of placements using the extreme ends of the Likert scale as shown in Fig. 1, and a position in the political landscape at the respective edges as shown in Fig. 2. Walgrave10 reminds us that the selection of statements in a VAA can greatly impact its results, cautioning against the interpretation of the candidates’ positions in the political landscape as an ideal point on an ideological dimension. We include these figures simply as a point of reference.
User data
This section starts by outlining the VAAs’ dissemination strategy during the election campaigns. It then details the data-generating process, after which it discusses the reach and make-up of the data.
The VAAs were primarily disseminated through major national media outlets in selected countries. Some of these outlets commissioned the tool directly from the developing research institute, while others translated the tool and received permission to host it on their platforms through an embedded iframe. We consider the outlets that incorporated the tool into their own platforms to be media partners, as detailed in Table 1. Figure 3 illustrates how the tool was embedded via an iframe within an article by HLN26. All participating media partners integrated the tool in a similar manner, except The Hindu who featured the tool as a standalone popup article on its homepage in 2024. Several partners also published follow-up articles analyzing aggregated user responses, sometimes embedding the tool again within their coverage. In addition, there were media outlets that simply reported about the tool that was available in their country, without the inclusion of such an iframe. While the media partners tended to be the largest source of traffic, the German tool in 2020 for instance attracted substantial traffic from outlets like T-Online who simply reported about the VAA. In 2024, the project was awarded the University of Amsterdam’s Alumni Impact Prize, which provided additional funding to support dissemination beyond European media. The tools for the 2020 and 2024 elections remain accessible via usa2020.electioncompass.org and usa2024.electioncompass.org, respectively.
Users go through a series of pages to arrive at the result page of the tool, as indicated in Fig. 4. After clicking ‘Start’ on the start page, users move through a series of demographic questions. These include age, gender, education, geographic location, and, in some 2024 cases, which party the user voted for in their last national election. Users proceed to the issue statements, which are shown one-by-one with the Likert scale below them. Users can return to update their answers, indicate that they have no opinion on the issue at hand, or skip the issue altogether. At this point, the 2020 and 2024 tools diverge slightly. The 2024 tools display a series of feeling thermometer items on 11-point scales, asking how favorably users feel about the presidential candidates, the two political parties, the American people, and the governments of the United States, China, and Russia. These questions are also optional, and users are able to select either ‘No opinion’ or ‘Prefer not to say’ for each of them. The software did not allow randomizing the order of these items. The 2024 tools ask another set of attitudinal questions before displaying the results of the tool. The attitude items ask the user about their economic and cultural attitudes, presidential candidate voting preference, and attitude toward the United States. These items are optional in both VAAs. Similar items were asked in 2020 after displaying the results, leading to lower item response rates. The 2020 tools move straight into the result pages after users answered the 26 issue statements. The 2020 tool shows three result pages: a ranking of both candidates, indicating a percentage agreement score; a two-dimensional landscape, akin to the ones shown in Fig. 2, which now includes the position of the user; and a page which details both candidates’ positions as well as the sources that underpin these positions. The 2024 tool shows only the first and last of these result pages, leaving out the political landscape due to the limited number of included items pertaining to the economic dimension. Users have the option to filter their results by the broad issue statement themes. All of these different pages in the tools collect users’ responses, producing a rich public opinion dataset.
Note that the 2020 French tool, disseminated by 20 Minutes, linked immediately to the issue statement page so that a large number of users did not see the demographic background items. To account for this, the same background items were added in the attitude items shown at the end of the tool, and subsequently entered into the demographics data.
In addition to the data entered by users, or calculated on their responses, the tools collect certain fields automatically.
We discuss some descriptives of interest below to provide a sense of the data. First, in Fig. 5, we show user numbers by day for all of the tools. It is clear that tools attract most users on and immediately after the day that their respective media partners launched the tool. Occasional peaks in the run-up to the election tend to be attributable to additional media attention, in nearly all instances by the disseminating outlet itself. In the days immediately before and on election day itself, numbers also increase relative to previous days. We include data for a number of days after the election to allow for analysis of changing patterns.
Second, while most responses originated from the countries in which the VAAs were disseminated, responses came in from all around the world. Figure 6 displays which countries’ responses were registered from (i.e., not where users indicated to live).
Third, we provide some indication of the representativeness of the samples. Since users were not invited to participate, but rather self-selected into participation with no intervention on the part of the authors aside from the collaboration with the media partners, the samples display biases commonly seen in nonprobability samples, particularly those stemming from VAAs3,4,6. As Fig. 7 shows, all tools know a gender bias with men being overrepresented (the black dot indicates the 50% proportion). The Male and Female bars are rescaled to 100% for straightforward interpretation, whereas the negative grey bars indicate the total proportion of respondents answering “Other,” “Prefer not to say,” or those with missing gender data. In addition, the tools were overwhelmingly consulted by users who enjoyed higher education. The dots on the bottom facet indicate the education levels in the population of the respective countries, as drawn from UNESCO data. There is thus an overrepresentation of higher-educated users and male users.
All data were collected in accordance with the European General Data Protection Regulation (GDPR) as certified by Privacy Certified. No personally identifiable data was collected. Data collection was approved by the VU University Amsterdam (2024-10-15-168). By commencing the use of the tools, all respondents consented to participation and agreed to the cookie and privacy policy (to be found at kieskompas.nl/nl/policies), which includes consent for anonymous data sharing.
Data Records
We describe 4 datasets and 3 other files that are made available on the Inter-university Consortium for Political and Social Research (ICPSR) database [238229]27. The first two datasets comprise the candidates’ placements on 29 statements in 2020 and 62 statements in 2024. The second two comprise the user data, reflecting 571,624 users in 2020 and 152,670 users in 2024. We also include a codebook for these four datasets. In addition, we include an overview of all statements between both election cycles, and which VAAs these statements were included in, as well as a file with code to replicate the tables, figures, and numbers in this manuscript, including code that follows our recommended usage notes.
The placement data includes two files, named ‘candidate_placements_2020.csv’ and ‘candidate_placements_2024.csv’. Each file is structured identically. The first column ‘statement_code’ has a unique identifier for the issue statement, followed by the second column ‘statement_text’. Note that these do not uniquely identify the row, which is uniquely identified in combination with the third column ‘candidate’, indicating the candidate being placed against a given statement. In 2020, these are Biden and Trump; in 2024, these are Trump and Harris. The following column ‘candidate_position’ indicates their assigned position. Subsequent columns contain the sources used to underpin the candidates’ position. Each position has up to three sources, each consisting of three columns: ‘source_1_origin’, ‘source_1_url’, and ‘source_1_excerpt’. Each row furthermore specifies whether we consider a statement to address a cultural or an economic issue, and which direction the statement is scaled under ‘statement_scaling’. Statement scaling takes values L for economic left, R for economic right, P for culturally liberal (progressive), and C for culturally conservative. Lastly, ‘in_vaa’ indicates whether a statement was included in the VAA, and, if so, in which one. Table 4 lists these variables.
In addition, we share two user data files, corresponding to users’ responses from the 2020 and 2024 surveys. These files are named ‘user_data_2020.csv’ and ‘user_data_2024.csv’ respectively. The files’ first column ‘tool’ identifies the tool source. The second column ‘token’ identifies the user with a unique alphanumerical combination. The column ‘language’ indicates the language in which the tool was completed, although users could also rely on applications like Google Translate to translate the content. Such translations would not be reflected in this column. The next two columns ‘date’ and ‘session_start’ indicate the day and the precise time that a user visited the tool. The ‘before_after_election’ converts the start time to whether the tool was filled out before or after the first election results came out. The next few variables pertain to how long users spent answering the tool in seconds, and how many seconds of those were spent on the issue statements. The data includes columns on what percentage of the statements were answered, and how many statements were answered in the same way, as well as what proportion of answered statements this entails. The next two variables indicate what device type was used to access the tool, and whether this was a touch device or not. The ‘continent’, ‘country’, and ‘subdivision’ variable contain automatically collected geographic data. As detailed under Usage Notes, we then include a number of variables that the user may find useful in filtering records. The following columns provide the entered demographic information by the respondent and are denoted with the prefix “bgq.” The next set of variables pertain to the user’s answers to the issue statements. These columns are named after the placement data’s ‘statement_code’ values. In 2020, the result variables follow. There are three columns with the prefix “pct”, which, based on the respondent’s answers to the issue statement questions, provide their percentage alignment with Joe Biden and Donald Trump, and then their highest agreement candidate. In addition, the 2020 data contains the user’s resulting ‘position_x’ and ‘position_y’. The 2024 data, before the percentage agreement result variables, includes the 11 feeling thermometer variables, denoted with the prefix “fts”. After that, eight columns describe the respondent’s attitudes toward a range of variables and are denoted with the prefix “att.” The 2020 data includes these additional attitudinal items after the result variables, to remain true to the ordering of the respective tools. We list all paradata and data selection criteria variables in Table 5, all user-entered data variables in Table 6, and finally all result variables in Table 7. Note that a more detailed version of the codebook can be found on ICPSR27.
Note that the 2024 tool for US audiences inadvertently got distributed in the Netherlands, and, therefore, many responses stem from the Netherlands as opposed to from the US. These responses can be recognized by the ‘country’ column.
Technical Validation
A number of checks can be conducted to validate our data. Starting with the placement data, we investigate the ideological positioning of the candidates. Logically, we would expect both Democratic candidates to be positioned economically left and culturally liberal with respect to Donald Trump. Figure 2 confirms that this is the case. In 2024, we followed a more rigorous coding process that had two coders place both candidates independently. They achieved an intercoder reliability of 0.78 (weighted Cohen’s kappa, ICC, and Krippendorf). Future studies may seek convergent validity e.g. through artificial intelligence coding estimates.
Second, we investigate the consistency of users’ answers. One way we might do that is by looking at the correlation between the tool’s results and users ideological self-placement. In 2020, we can compare with both the users’ percentage agreement scores with the candidates and with the users’ positions in the landscape. The correlation between left-right self-placement is 0.53 (p < 0.001) with ‘position_x’, −0.54 (p < 0.001) with agreement with Biden, and 0.54 (p < 0.001) with agreement with Trump. The correlation between conservative-liberal self-placement is 0.49 (p < 0.001) with ‘position_y’, 0.48 (p < 0.001) with agreement with Biden, and −0.48 (p < 0.001) with agreement with Trump. These patterns are highly similar for all 2020 tools separately as well. In 2024, we can only compare with the agreement scores. The correlation between left-right self-placement is −0.43 (p < 0.001) with agreement with Harris and 0.45 (p < 0.001) with agreement with Trump. The correlation between conservative-liberal self-placement is 0.43 (p < 0.001) with agreement with Harris, and −0.43 (p < 0.001) with agreement with Trump. Similarly, we would expect high correlations between Trump and Harris’ thermometer ratings and the percentage agreement with each. Note that the thermometers were only asked in 2024. The correlation between Trump’s thermometer rating and agreement with Trump is 0.69 (p < 0.001). For Harris, this correlation is 0.67 (p < 0.001). This provides some evidence that the outcome variables of the tools align with users’ ideologies.
Third, to ensure robustness of the results shown to users, we may investigate what proportion of users received results where both candidates are close to each other. Both in 2020 and in 2024, fewer than 1% of respondents received results where agreement with both candidates was equal. Precisely 5% of users in 2020 were shown results where the difference between the candidates was less than 5%. In 2024, 6.4% of users were shown results where the difference between the candidates was less than 5%. This provides evidence that for the overwhelming majority, statement selection did not determine the comparison between the candidates.
Fourth, the number of seconds that users spend answering the tool should never be smaller than the number of seconds they spend on the statements alone. This is true in all tools. On average, users spend 213 seconds on the tool, of which 147 seconds on the issue statements and 66 seconds elsewhere. These durations appear to pass a face validity test.
Lastly, due to the gamified nature of VAAs, we risk that users do not provide genuine answers at higher rates than they might in other surveys. In 2024, we asked users whether they filled out the tool while expressing their genuine views, trying to end up closer to either candidate, or just playing around. Of all respondents who answered this question, 89.5% say they were answering genuinely. Of course, for a user who’s answering randomly, this answer may not be so credible. However, we may investigate the consistency of these users’ answers. We would expect higher internal consistency on both dimensions for the genuine responses compared to the random responses. As expected, genuine answers have a Cronbach’s alpha of 0.44 on the economic dimension and of 0.85 on the cultural dimension, whereas random answers have a Cronbach’s alpha of 0.33 on the economic dimension and of 0.75 on the cultural dimension.
Usage Notes
To ensure the validity and reliability of the data being used, it is recommended that any records which do not meet any of the first three criteria shown in Table 8 be filtered out.
All of these criteria are reflected in their respective variables (see Table 4). The variable ‘crit_all’ combines the top three criteria as recommended, allowing for simple filtering of records. While the genuine response criterion (included in 2024 only) is not independent from the combined criteria (χ²(1) = 82, p < 0.0001), it performs below expectations substantively (Cohen’s Kappa = 0.02). Since the criterion does not feature in the 2020 data, it is not included in the ‘crit_all’ variable, but users may consider using it to filter records, particularly when investigating the 2024 data.
Data availability
The data described above is made available on the Inter-university Consortium for Political and Social Research (ICPSR) database [238229]27. It includes two datasets comprising the candidates’ placements on 29 statements in 2020 and 62 statements in 2024, as well as two user datasets covering 571,624 user records in 2020 and 152,670 user records in 2024. We also include a codebook for these four datasets. In addition, we include an overview of all statements between both election cycles, and which VAAs these statements were included in.
Code availability
Replication code and recommended code are available in the ‘code’ folder in the same location as the data files on the ICPSR database27.
References
Garzia, D. & Marschall, S. Matching Voters with Parties and Candidates: Voting Advice Applications in a Comparative Perspective. (ECPR Press 2014).
Krouwel, A., Holleman, B., Kamoen, N., van de Pol, J. & de Vreese, C. Effecten van online stemhulpen. (2018).
Van De Pol, J., Holleman, B., Kamoen, N., Krouwel, A. & De Vreese, C. Beyond Young, Highly Educated Males: A Typology of VAA Users. Journal of Information Technology & Politics 11, 397–411 (2014).
van de Pol, J., Kamoen, N., Krouwel, A., de Vreese, C. & Holleman, B. Same but different: A typology of Voting Advice Application users in first- and second-order elections. Acta Politica 54, 225–244 (2018).
Etienne, T. The persistent nature of differential participation in the nonprobability sampling process of a voting advice application generated panel. SSRN Scholarly Paper at https://papers.ssrn.com/abstract=4925197 (2023).
Albertsen, A. How do the Characteristics of Voting Advice Application Users Change Over Time? Evidence from the German Election Studies. German Politics 31, 399–419 (2022).
Andreadis, I., Wall, M. & Krouwel, A. Who are the users of voting advice applications? (2017).
Holleman, B., Oldenhuis, P. & Kamoen, N. Issueframing als een vorm van equivalence framing: Het effect van progressieve versus conservatieve introductieteksten op gerapporteerde attitudes in politieke stemhulpen. Tijdschrift voor Taalbeheersing 42, 279–308 (2020).
Kamoen, N. & Holleman, B. I don’t get it. Response difficulties in answering political attitude statements in voting advice applications. Survey Research Methods 11, 125–140 (2017).
Walgrave, S., Nuytemans, M. & Pepermans, K. Voting advice applications and the effect of statement selection. West European Politics 32, 1161–1180 (2009).
Louwerse, T. & Rosema, M. The design effects of voting advice applications: Comparing methods of calculating matches. Acta Polit 49, 286–312 (2014).
Mendez, F. Matching voters with political parties and candidates: an empirical test of four algorithms. International Journal of Electronic Governance 5, 264–278 (2012).
Talonen, J. & Sulkava, M. Analyzing Parliamentary Elections Based on Voting Advice Application Data. in Advances in Intelligent Data Analysis X 340–351, https://doi.org/10.1007/978-3-642-24800-9_32 (Springer, Berlin, Heidelberg, 2011).
Gemenis, K. Estimating parties’ policy positions through voting advice applications: Some methodological considerations. Acta Politica 48, 268–295 (2013).
Katsanidou, A. & Lefkofridi, Z. Citizen Representation in the European Union: Policy Congruence in the 2009 EP Election. in An Audit of Democracy in the EU (eds. Eijk, C. van der et al.) 153–174 (European University Institute, Florence, 2013).
Bethlehem, J. How Accurate Are Self-Selection Web Surveys. Statistics Netherlands 1–24 (2008).
Bethlehem, J. Selection Bias in Web Surveys. International Statistical Review 78, 161–188 (2010).
Antonucci, L., Horvath, L., Kutiyski, Y. & Krouwel, A. The malaise of the squeezed middle: Challenging the narrative of the ‘left behind’ Brexiter. Competition & Change 21, 211–229 (2017).
Van Bavel, J. J. et al. National identity predicts public health support during a global pandemic. Nature Communications 13, 1–14 (2022).
Van Prooijen, J.-W., Etienne, T., Kutiyski, Y. & Krouwel, A. Conspiracy Beliefs Prospectively Predict Health Behavior and Well-being during a Pandemic. Psychological Medicine 1–25 (2021).
Toshkov, D. & Romeijn, J. How to estimate the policy preferences of party supporters: Disaggregating data from voting advice applications versus modeling survey responses. Electoral Studies 74, 102403 (2021).
Krouwel, A., Vitiello, T. & Wall, M. The practicalities of issuing vote advice: A new methodology for profiling and matching. International Journal of Electronic Governance 5, 223–243 (2012).
Krouwel, A. & van Elfrinkhof, A. Combining strengths of methods of party positioning to counter their weaknesses: the development of a new methodology to calibrate parties on issues and ideological dimensions. Quality & Quantity 48, 1455–1472 (2013).
Dillman, D. A., Smyth, J. D. & M Christian, L. Internet, Phone, Mail, and Mixed-Mode Surveys. (Wiley, Hoboken, NJ, 2014).
Dias, N. Do Citizens’ Policy Opinions Match Their Values? Applying a New Test of Vertical Constraint. in (Chicago, 2024).
IHB. Doe de test: hoeveel procent Trump of Harris ben jij? hln.be (2024).
Etienne, T. W. et al. A Global Voting Advice Application for the US Election Aimed at Foreign Audiences. Inter-university Consortium for Political and Social Research (ICPSR) https://doi.org/10.3886/E238229V3 (2025).
Rovny, J. et al. 25 Years of Political Party Positions in Europe: The Chapel Hill Expert Survey, 1999–2024 (2024).
Lehmann, P. et al. The Manifesto Data Collection. Manifesto Project (MRG/CMP/MARPOR). Wissenschaftszentrum Berlin für Sozialforschung (WZB) https://doi.org/10.25522/manifesto.mpds.2024a (2024).
Norris, P. Global Party Survey, 2019. Harvard Dataverse https://doi.org/10.7910/DVN/WMGTNS (2020).
Coppedge, M. et al. V-Dem [Country-Year/Country-Date] Dataset v15. Varieties of Democracy (V-Dem) Project (2025).
Acknowledgements
We are grateful to the support of the Kieskompas – Election Compass research institute for the use of their software and the permission to publish this data. We thank all participating media outlets, in particular HLN and AD who were the first two outlets to sign up for this project. We thank the University of Pennsylvania and the University of Amsterdam for their financial support. We also thank our colleagues who translated the tools or verified translations, Thomas Vitiello and Mihnea Stoica, and Aswin Punathambekar for his advice on the 2024 tool.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Contributions
Tom W. Etienne was the project lead for the 2020 and 2024 VAA development. In 2020, statement development and candidate placements were supported by Richard. A. Furstein, Yordan Kutiyski, Jeroen van Lindert, and Andrew Pasquier. Jan Philipp Thomeczek translated and disseminated the tool in Germany. In 2024, statement development and candidate placement were taken on by Lily Markis-McLean and Neil V. Bennett. Oscar Moreda-Laguna provided logistical support, and Andre P.M. Krouwel provided academic support.
Corresponding author
Ethics declarations
Competing interests
Several authors are currently or have in the past been employed or affiliated with the Kieskompas - Election Compass research institute. Andre Krouwel and Jeroen van Lindert hold shares in the research institute.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Etienne, T.W., Bennett, N.V., Markis-McLean, L. et al. A Global Voting Advice Application for the US Election Aimed at Foreign Audiences. Sci Data 12, 1906 (2025). https://doi.org/10.1038/s41597-025-06333-6
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41597-025-06333-6









