Hostname: page-component-857557d7f7-zntvd Total loading time: 0 Render date: 2025-12-06T05:42:52.004Z Has data issue: false hasContentIssue false

Measuring Politicians’ Public Personality Traits Using Computational Text Analysis: A Multimethod Feasibility Study for Agency and Communion

Published online by Cambridge University Press:  02 December 2025

Lukas Birkenmaier*
Affiliation:
GESIS - Leibniz-Institut fur Sozialwissenschaften eV, Germany
Clemens Lechner
Affiliation:
GESIS - Leibniz-Institut fur Sozialwissenschaften eV, Germany
*
Corresponding author: Lukas Birkenmaier; Email: lukas.birkenmaier@outlook.de
Rights & Permissions [Opens in a new window]

Abstract

Citizens’ opinions about politicians are shaped by their perceptions of politicians’ personalities, characters, and traits. While prior research has investigated the traits voters value in politicians, less attention has been given to the traits politicians project in their public communication. This may stem from challenges in defining politicians’ public personality traits and measuring them at scale using computational text analysis. To address this challenge, we propose a computational approach that builds on public statements (personality cues) to infer politicians’ personalities from textual data. To do so, we operationalize two key political traits—agency and communion—using a theory-driven, domain-specific framework. We then compare various computational text analysis methods for extracting these traits from a large corpus of politicians’ parliamentary speeches, social media posts, and interviews. We validate our approach using a comprehensive set of human-labeled data, functional tests, and analyses of how prominently personality traits appear in the statements of German politicians and in the 2024 U.S. presidential debate between Donald Trump and Kamala Harris. Our findings indicate that prompting based techniques, particularly those leveraging advanced models such as DeepSeek-V3, outperform supervised and semisupervised methods. These results point to promising directions for advancing political psychology.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of The Society for Political Methodology

1 Introduction

The way citizens judge politicians is influenced by how they evaluate politicians’ character or personality traits (Bittner and Peterson Reference Bittner and Peterson2018; Hayes Reference Hayes2009; Laustsen and Bor Reference Laustsen and Bor2017). A more favorable evaluation of a politician’s personality increases the likelihood of voting for them and contributes to the formation of positive partisan attitudes (Funk Reference Funk1997; Miller, Wattenberg, and Malanchuk Reference Miller, Wattenberg and Malanchuk1986). The abiding trends toward greater personalization of politics renders voters’ perceptions of politicians’ personality even more important, as individual political actors become more prominent at the expense of collective actors and institutions such as political parties (Pedersen and Rahat Reference Pedersen and Rahat2021; Rahat and Kenig Reference Rahat and Kenig2018).

While a vast literature has studied which personality traits voters look for in politicians (see Van Aelst, Sheafer, and Stanyer Reference Van Aelst, Sheafer and Stanyer2012 for an overview), less is known about the ways in which politicians present their personality in public communication, such as through speeches or social media posts. Previous studies have explored various aspects of politicians’ strategic self-presentation, including their style of representation (Fenno Reference Fenno2003; Gulati Reference Gulati2004; Schürmann and Stier Reference Schürmann and Stier2022) or their private lives (Colliander et al. Reference Colliander, Marder, Falkman, Madestam, Modig and Sagfossen2017; Perloff Reference Perloff2013; Stanyer Reference Stanyer2008; Van Zoonen Reference Van Zoonen2005). However, little research has specifically examined how politicians use public communication to convey personality traits to the public, either strategically or inadvertently. One reason for this gap may be the difficulty of measuring a politician’s public personality when the available information is often fuzzy and ambiguous.

Leveraging recent advances in computational text analysis and political psychology, our study establishes the feasibility of inferring politicians’ public personality from purely verbal personality cues. Although public personality can also be conveyed through nonverbal (e.g., posture, gestures) and paraverbal (e.g., prosodic features) cues, verbal cues (i.e., what politicians say) stand out as both informative (Chen, Qiu, and Ho Reference Chen, Qiu and Ho2020) and widely available for analyzing political figures in public communication. The core methodological contribution of our study is a computational text analysis workflow that detects verbal personality cues in political communication and infers politicians’ public personality traits from these cues.

We build upon a validated personality framework and attendant survey inventory to operationalize two key traits in political psychology: agency and communion. We then apply different methods to automatically infer them at scale in a text corpus comprising social media posts, political speeches, and political interviews. We systematically compare multiple modeling strategies and validate our approach within a structured evaluation framework. Our results suggest that prompt-based approaches using state-of-the-art models, such as DeepSeek-V3 and GPT-4o, show the strongest evidence of measurement validity and are best suited to identify relevant personality cues.

To support transparency and future research, we provide all trained models, code, and data openly.Footnote 1

2 Literature

2.1 Politicians’ Personality Traits Matter

Politicians’ personality plays a central role in explaining political actions, such as partisan attitudes or voting behavior (Best Reference Best2011; Federico and Malka Reference Federico and Malka2018; Schoen and Schumann Reference Schoen, Schumann, Falter, Gabriel and Weßels2005). While politicians and their characteristics have always played a key role in politics, there are growing claims that a new, increasingly personalized political culture of “candidate-centered politics” has emerged (McAllister Reference McAllister, Dalton and Klingemann2007; Pedersen and Rahat Reference Pedersen and Rahat2021). This shift is attributed to a decline in the traditional emotional connections between voters and political parties (Mair, Reference Mair2005) and the transformative impact of a media landscape that prominently includes decentralized social media (Stier et al. Reference Stier, Bleier, Lietz and Strohmaier2018). The growing interest in the characteristics of politicians has led to an increased focus on analyzing politicians’ personalities.

For example, previous research has examined the role of politicians’ personality as a means to understand their ideological orientations and political behavior (George and George Reference George and George1998; Walker, Schafer, and Young Reference Walker, Schafer and Young1998).Footnote 2 This strand of research indicates that conservative politicians often show lower levels of agreeableness and openness but higher levels of conscientiousness, whereas the link to other personality traits is less clear (Hanania Reference Hanania2017; Schumacher and Zettler Reference Schumacher and Zettler2019). Evidence also suggests that politicians, compared to the general population, show higher levels of agreeableness, emotional stability, and openness (Caprara et al. Reference Caprara, Barbaranelli, Consiglio, Picconi and Zimbardo2003). Likewise, personality, particularly higher levels of extraversion and lower levels of agreeableness, is related to higher in-office performance (Joly, Soroka, and Loewen Reference Joly, Soroka and Loewen2019; Lilienfeld et al. Reference Lilienfeld, Waldman, Landfield, Watts, Rubenzer and Faschingbauer2012; Watts et al. Reference Watts2013), media visibility (Amsalem et al. Reference Amsalem, Fogel-Dror, Shenhav and Sheafer2020), or electoral success (Nai Reference Nai2019; Pillai et al. Reference Pillai, Williams, Lowe and Jung2003; Scott and Medeiros Reference Scott and Medeiros2020).

Naturally, voter preferences also display influences of personality-based judgments. For example, voters appear to consistently prefer characteristics such as competence, integrity, or energy in political figures (Aichholzer and Willmann Reference Aichholzer and Willmann2020; Mondak Reference Mondak1995). Caprara and Zimbardo (Reference Caprara and Zimbardo2004) further elaborate that voters’ evaluation of politicians depends on how closely politicians resonate with the voters’ own traits and values. This can manifest through emotional attachments (likability heuristic) or the attribution of positive characteristics (dispositional heuristics). Cross-national studies confirm varying preferences for traits such as emotional stability, extraversion, and conscientiousness in ideal politicians (Aichholzer and Willmann Reference Aichholzer and Willmann2020).

2.2 Politicians’ Public Personality Is Measured Through Cues

In nearly all cases, voters must rely on secondary sources (e.g., media) to assess a politician’s personality. Therefore, voters’ judgments are exclusively based on the public persona projected by the politician. Figure 1 outlines this process in an adaptation of the lens model proposed by Brunswik (Reference Brunswik1952).

Figure 1 Conceptual model (adapted from Brunswik, Reference Brunswik1952).

The key message of the model is that voters and researchers alike never have direct access to a politician’s intrinsic personality. Rather, they have to infer politicians’ personality from (i.e., through the lens of) observable personality cues ${X}_1$ ${X}_n$ communicated by the politician. These cues, whether verbal, such as statements in speeches or debates, or nonverbal, such as gestures, posture, or facial expressions, are the basis upon which observers form their judgments of a politician’s personality.

The observable personality cues ${X}_1$ ${X}_n$ are influenced by two main factors: a politician’s intrinsic personality ( ${Y}_{IPT}$ ) and their strategic self-presentation ( ${Y}_S$ ), in which they deliberately emphasize or suppress certain traits to appeal to voters (Fenno Reference Fenno2003; Goffman Reference Goffman2016). The relative influence of intrinsic dispositions and strategic behavior in any given cue cannot be determined. However, this uncertainty is not a limitation when the focus is, as in our study, precisely on automating the personality judgments that the public (i.e., voters or news media) would also form based on the cues.

The challenge for inferring politicians’ personality through automated text analyses, however, is twofold: (1) to identify valid cues ${X}_1$ ${X}_n$ that can serve as indicators for a given personality trait (cue validity), and (2) to correctly utilize these cues to form judgments of that trait (cue utilization).

2.3 Personality Cues Are Inherently Ambiguous

Identifying personality cues and utilizing them to form coherent trait judgments is inherently challenging for two main reasons.

2.3.1 Cue Validity

First, personality traits themselves are not sharply defined or neatly separable—they have the nature of fuzzy concepts (Thielmann and Hilbig Reference Thielmann and Hilbig2019). As such, traits like extraversion and agreeableness do not always have clear-cut boundaries, making it difficult to determine where one ends and another begins (Zettler et al. Reference Zettler, Thielmann, Hilbig and Moshagen2020). In addition, not all traits are equally observable to begin with. Traits such as extraversion often have relatively clear behavioral cues—like a politician’s public speaking style or social engagement—whereas others, such as openness or honesty, are harder to detect.

2.3.2 Cue Utilization

Second, the observable manifestations of these traits—in our case the personality cues ${X}_1$ ${X}_n$ —are themselves ambiguous. A single behavior can be driven by multiple traits (equifinality), and conversely, a single trait can result in a wide range of behaviors (multifinality). This means that very few behaviors can be considered “pure” indicators of a single trait; instead, personality cues tend to remain ambiguous stimuli open to interpretation. In political contexts, these ambiguities are amplified by the viewer’s own perspective. Partisan or cultural biases can shape how observers interpret the same cue—what one supporter sees as “bold” or “charismatic,” an opponent might perceive as “reckless” or “showboating.” This means that not all observers will use the same available cues in the same way, let alone use them correctly, to infer personality traits.

In sum, personality as a construct is both conceptually and empirically fuzzy, and its behavioral indicators are often ambiguous and context dependent. It is not always straightforward to identify valid, let alone pure, cues of each given personality trait (cue validity). It is also challenging for observers, whether human or computers, to utilize the right cues to correctly infer a given personality trait (cue utilization). This makes the reliable measurement of personality in public discourse particularly demanding.

2.4 How Can We Utilize Personality Cues?

Due to these challenges, methods for measuring politicians’ personalities based on personality cues face considerable limitations (McDonald Reference McDonald2008). Broadly speaking, these methods can be categorized into self-report and observer-report approaches.

Self-report approaches include participants being asked to fill out surveys (Caprara et al. Reference Caprara, Barbaranelli, Consiglio, Picconi and Zimbardo2003; Joly, Soroka, and Loewen Reference Joly, Soroka and Loewen2019; Maier et al. Reference Maier, Oschatz, Stier and Zettler2023; Nørgaard and Klemmensen Reference Nørgaard and Klemmensen2019; Schumacher and Zettler Reference Schumacher and Zettler2019).Footnote 3 However, politicians—especially high-profile politicians—are a hard-to-survey population, and surveys among politicians often suffer from low response rates, which not only make data access challenging but may also lead to biased response patterns. Likewise, time limitations frequently lead researchers to use brief personality inventories, which may fail to fully represent the complexity of a politician’s personality profile. Also, fundamentally measuring politicians’ intrinsic personality in this way is not the same thing as measuring their personality from the observable cues based on which the public would form their personality judgments.

On the other hand, observer-report approaches typically rely on observer ratings derived from various forms of observational data, such as everyday interactions or digital traces from social media. In the study of politicians’ personalities, this can be done using ratings by experts or the general public (de Vries and van Prooijen Reference Vries and van Prooijen2019; Rubenzer, Faschingbauer, and Ones Reference Rubenzer, Faschingbauer and Ones2000; Visser, Book, and Volk Reference Visser, Book and Volk2017; Wright and Tomlinson Reference Wright and Tomlinson2018; Nai and Maier Reference Nai and Maier2024). Although these evaluations are explicitly based on observable personality cues ( ${X}_1$ ${X}_n$ ), they can be skewed by the raters’ political beliefs, and these raters have, by definition, only a partial and biased view of the politicians they are evaluating (Wright and Tomlinson Reference Wright and Tomlinson2018). In addition, the interpretation and aggregation of personality cues into traits by raters often hides the direct connection between specific cues and the final evaluation, hampering the observation of short-term changes and specific settings.

2.5 The Need for Computational Approaches

In this article, we therefore argue that computational text analysis provides a valuable yet underutilized tool for measuring politicians’ public personality. Generally, political communication serves as the primary medium for self-presentation, with every statement made by a politician potentially being a personality cue that voters perceive. In the past, research has aimed to study political self-presentation using manual content analysis (Hermann Reference Hermann1980; Winter Reference Winter2005). However, content analysis usually entails a targeted, case-study-specific angle for politicians like Donald Trump and Hillary Clinton (Lee and Lim Reference Lee and Lim2016) or George Bush and Mikhail Gorbachev (Winter et al. Reference Winter, Hermann, Weintraub and Walker1991). Therefore, while manual content analysis likely provides the most valid research approach in a case-study design, it is unnecessarily effortful for analyzing politicians public images on a large scale, given that automated personality prediction from behavioral data has demonstrated performance comparable to traditional approaches (see the meta-analysis by Azucar et al., Reference Azucar, Marengo and Settanni2018).

A notable exception in the use of automated methods is the study by Ramey et al. (Reference Ramey, Klingler and Hollibaugh2019), who used computational techniques to measure the personalities of U.S. Congress members based on their floor speeches. Their approach builds on a rich literature that uses tools like Linguistic Inquiry and Word Count (LIWC) to capture linguistic features—such as the use of second-person pronouns, punctuation, or word length—and link them to personality traits (Chen, Qiu, and Ho Reference Chen, Qiu and Ho2020; Park et al. Reference Park2015; Mairesse et al. Reference Mairesse, Walker, Mehl and Moore2007). Using the pretrained Support Vector Machines for Regression models from Mairesse et al. (Reference Mairesse, Walker, Mehl and Moore2007) and applying it to U.S. Congress floor speeches, they found that their text-based measures were correlated with survey response rates for members of Congress.

While the work of Ramey et al. represents a significant contribution by introducing a computational method to extract politicians’ personality cues, it also has limitations. Their use of a pretrained support vector machine lacks validation against human-annotated data and appears outdated given advances in large language models (Laurer et al. Reference Laurer, van Atteveldt, Casas and Welbers2024; Widmann and Wich Reference Widmann and Wich2022). Moreover, the Big Five traits, though common in personality research, offer little domain-specific nuance for political contexts.

Building on the work of Ramey et al., we, therefore investigate whether computational methods can be used to develop valid, domain-specific automated measures of politicians’ public personalities based on the personality cues they exhibit.

3 Method

3.1 Data

We compiled a corpus comprising a diverse set of public statements from leading German politicians covering the period from 2017 to 2024, encompassing two full legislative terms of the German Bundestag. Our definition of “leading politicians” includes those who served as leaders or co-leaders of both legislative factions and political parties, covering all six parties represented in the German Bundestag. For these 46 politicians (see Appendix A of the Supplementary Material), we collected data from three sources commonly used to study politicians’ public communication: interviews, social media posts, and parliamentary speeches.

3.1.1 Interviews

First, we rely on a novel dataset comprising 99 interviews with political leaders broadcasted in five different media formats (Birkenmaier, Sieber, and Bergstein, Reference Birkenmaier, Laureen and Felix2025). Interview statements provide one of the most natural and authentic sources of politicians’ interactions and self-presentation to the public, ideally lending themselves to capture the public personality of politicians (Schoonvelde, Schumacher, and Bakker Reference Schoonvelde, Schumacher and Bakker2019).

3.1.2 Social Media Posts

Second, we rely on social media data as another highly popular resource for politician’s public communication. The initial data consist of a subsample of 300,000 Facebook posts from a raw corpus of 6.7 million politicians’ posts utilized in Birkenmaier, Sältzer, and Wurthmann (Reference Birkenmaier, Wurthmann and Sältzer2025).

3.1.3 Parliamentary Speeches

Finally, we rely on a corpus of politicians’ speeches in the German Bundestag. We utilize a subset of German MP speeches provided by the Parlspeech database, which provides transcripts of parliamentary speeches of nine representative democracies (Rauh and Schwalbach Reference Rauh and Schwalbach2020). We used only the German subset for our present analyses.

Across all data sources, we defined individual sentences and statements uttered or written by politicians as our unit of analysis. Sentences represent personality cues at a granular level that can be detected and potentially aggregated across texts. Because our primary focus is on detecting verbal personality cues in political communication and inferring politicians’ public personality traits from these cues in text rather than analyzing explanatory factors influencing changes or differences in the public personality, we draw a random sample of 15,000 sentences from each data source as our main corpus.

3.2 Measurement

3.2.1 Workflow

Figure 2 illustrates our workflow. As detailed next, we begin with manual coding to develop and validate a codebook and create a set of labeled data. We then employ computational text analysis using different models and measurement strategies and assess the validity of our measures using a comprehensive validation framework (Birkenmaier, Wagner, and Lechner Reference Birkenmaier, Wagner and Lechner2024).Footnote 4

Figure 2 Workflow.

3.2.1.1 Trait Selection.

To select personality traits that can be measured using computational text analysis, we focus on two traits that are widely used in personality and political science research: agency and communion (Bakan Reference Bakan1966). These so-called “meta-traits” provide a higher order structure that captures common variance among lower order traits, including the Big Five, and consistently emerge across diverse personality models (e.g., Partsch, Bluemke, and Lechner Reference Partsch, Bluemke and Lechner2022; Saucier et al. Reference Saucier2014). Adopting the terminology of Hogan (Reference Hogan1982), agency can be thought of as the drive to “get ahead,” whereas communion reflects the drive to “get along.” These dimensions constitute key dimensions of person perception and important behavioral determinants in both personal and political environments (Entringer, Gebauer, and Paulhus Reference Entringer, Gebauer and Paulhus2022).

To guide our operationalization of agency and communion, we rely on a validated framework and survey inventory to assess politicians’ personalities proposed by Lechner et al. (Reference Lechner, Aichholzer, Birkenmaier and Bluemke2025). This inventory comprises 15 domain-specific personality facets specifically tailored to assessing politicians’ personality. Compared to global and general personality frameworks such as the Big Five (Costa Jr. and McCrae Reference Costa and McCrae1995), the dimensions in the Framework by Lechner et al. offer a validated, domain-specific framework for politicians’ personality traits, providing both a clear definition and a conceptual basis for subsequent computational text analysis. In our analysis, we focus on the two prototypical traits from their framework that capture the essence of agency and communion, respectively. This keeps our method interpretable, theoretically grounded, and capable of focusing on the validation of measuring these traits using personality cues in textual data.

Agency: In the Lechner et al. inventory, the trait Assertiveness & Will to Take Control is the key trait representing agency. It reflects the willingness to be a leader as well as to achieve power and one’s assertiveness. The items for this trait characterize a politician who is “power-hungry, claims leadership,” “assertive, dominant,” and “a born leader.”

Communion: In the Lechner et al. inventory, Empathy & Compassion constitutes the key trait representing communion.Footnote 5 It reflects the ability to express and act on feelings of empathy and compassion toward others. The items for this trait characterize a politician who “is empathetic, shows genuine compassion,” “understands the needs of disadvantaged people,” and “is compassionate, has a soft heart.”Footnote 6

3.2.1.2 Trait Conceptualization.

For each trait, we define both weak and strong cues that reflect varying degrees of explicitness in the signaling of personality cues. Personality cues can differ in both strength and relevance, as they vary in how clearly and accurately they reflect the traits being judged in real-world context (Funder Reference Funder1995). In our case, strong cues refer to clear and direct statements that contain self-disclosure or direct references to a personality trait ascribed to oneself (e.g., “I am a very stable genius”), whereas weak cues are more ambiguous or indirect and might be more open to subjective interpretation by the receiver of the cue. Agency, for example, can be explicitly signaled through strong statements such as “As the leader of this party, I will ensure that this policy is implemented, regardless of who stands in the way.” In contrast, weaker agency cues might include phrases that can potentially be subject to human interpretation, such as “We will push this policy forward and ensure its implementation.” While the latter statement suggests a call to action, it is much more passive and does not clarify whether the action entails the speaker actively overcoming resistance.

3.2.1.3 Codebook Development and Validation.

We developed the codebook in multiple steps. Initially, we only coded a subset of sentences containing elements of personal attributes using self-referential words as filtering keywords (e.g., “I,” “we,” “mine”; see Appendix C of the Supplementary Material). Coders annotated these sentences while providing continuous feedback to refine the codebook. Once coders achieved a consensus on trait assignments, we expanded the coding to a random selection of sentences from the full dataset to improve generalizability. Ultimately, a team of three annotators (i.e., the first author and two research assistants) annotated 1,000 sentences. We computed Krippendorff’s alpha as a measure of multilabel agreement, with values above α > .70 deemed as acceptable (Krippendorff Reference Krippendorff2018). After demonstrating sufficient agreement on the 1,000 sentences (see Section 4), we proceeded to generate a human-labeled corpus containing at least 200 examples per class (“agency,” “communion,” or “none”) for both weak and strong cues that we can use to compare our computational scores against (see Appendix D of the Supplementary Material for the final codebook).

3.2.2 Computational Analysis

With the growing application of large language models (LLMs) in the social sciences, there is an ongoing debate concerning the best strategies for computational text classification (Gilardi, Alizadeh, and Kubli Reference Gilardi, Alizadeh and Kubli2023; Ziems et al. Reference Ziems, Held, Shaikh, Chen, Zhang and Yang2024). Therefore, in this article, we systematically test different strategies and computational models to determine the best-performing architecture to detect valid personality cues in textual data.

For all models, we test their performance on different dataset variants. In the first dataset variant, referred to as strong, we only include examples of strong cues for agency and communion. In the second variant, referred to as weak + strong, we include an equal proportion of weak and strong cues for agency and communion, representing a more inclusive (but potentially less clear-cut) definition of personality trait expressions. To address class imbalance, we assess performance on both a balanced dataset, where the three classes (“agency,” “communion,” and “none”) are equally represented, and an unbalanced dataset, where “none” appears twice as often as the other classes.

We compare three major strategies across multiple model architectures (see Figure 2) and assess performance on different dataset variants. All hyperparameters and computational resources used in our experiments are documented in Appendices E and F of the Supplementary Material.

For the supervised strategy (Strategy 1), we follow the approach by Ramey et al. (Reference Ramey, Klingler and Hollibaugh2019) and train a Support Vector Machine (SVM) as a baseline model on the human-annotated data. SVMs are fast and computationally efficient, as they are trained on a bag-of-words representation—a simple encoding of word frequency that lacks contextual information. This makes them particularly suited for detecting direct lexical cues (e.g., explicit words connected to a personality traits), but less effective at picking up ambiguous or implicit personality cues. We compute fivefold cross-validation, meaning the data are split into five parts and the model is trained and tested five times, each time holding out a different part for testing, using a train-test split of 80/20.

For fine-tuning a pretrained LLM (Strategy 2), we rely on a 279M parameters XLM-RoBERTa (base) model (Conneau et al. Reference Conneau2020) using common hyperparameters. As a transformer-based model, XLM-RoBERTa captures contextual relationships between words, and fine-tuning it on our annotated data allows it to adapt to domain-specific language, making it well-suited for detecting implicit or contextually embedded personality cues. We compute threefold cross-validation, using a train-test split of 80/20.

For the prompting approach (Strategy 3), we prompt three different generative instruction tuned LLM, following current research exploring generative models for classification (Gilardi, Alizadeh, and Kubli Reference Gilardi, Alizadeh and Kubli2023; Ziems et al. Reference Ziems, Held, Shaikh, Chen, Zhang and Yang2024). Generative models are particularly effective at detecting subtle or fuzzy personality signals without additional training, as they draw on broad pretraining and are optimized to interpret tasks from instructions alone. We rely on three models: a proprietary instruction tuned GPT-4o, and two open-weight models Llama 3-8B (Touvron et al. Reference Touvron2023) and DeepSeek-V3 (DeepSeek-AI et al. Reference DeepSeek-AI2024). For each model, we utilize different prompting templates (see Appendix G of the Supplementary Material) tailored to dataset variants featuring either exclusively strong or a combination of strong + weak cues, as outlined in our codebook. Additionally, our approach distinguishes between zero-shot and few-shot settings, with the latter including class-specific examples. Each prompt directs the respective model to classify politicians’ statements into one of three categories (“agency,” “communion,” and “none”) and to provide a brief, one-sentence explanation for its decision in a structured format.

To mitigate the effects of randomness in training and prompting, we run each of our models with three different random seeds and report the macroaveraged F1 performance across runs, which reflects the harmonic mean of precision and recall computed separately for each class and averaged to avoid bias toward dominant categories.

To validate our measures, we take a systematic approach to validation and rely on the ValiText framework (Birkenmaier et al., Reference Birkenmaier, Wagner and Lechner2024). Appendix H of the Supplementary Material outlines our validation strategy in detail, and Appendix I of the Supplementary Material contains the checklist templates documenting each step in the validation process.

4 Results

4.1 Substantive Validation

As outlined in our method section, we employ a domain-specific personality framework and attendant survey inventory as the conceptual basis for operationalizing personality cues. As detailed in Lechner et al. (Reference Lechner, Aichholzer, Birkenmaier and Bluemke2025), the substantive validity of this personality framework was ensured during its development process through multiple means, including an analysis of extant work on personality assessment in general and politicians’ personality traits more specifically; input and reviews by experts from psychometrics and political science; as well as extensive tests of structural and external validity of the resulting inventory in multiple samples of voters.

We then chose the prototypical traits of the framework to operationalize personality cues for agency and communion. To do so, we developed a codebook and conducted manual coding to evaluate whether human annotators could reliably identify corresponding personality cues. Our analysis demonstrated sufficient agreement, with Krippendorff’s alpha of .73 (strong) and .72 (weak + strong) for the overall identification of the presence of personality cues, and agreement ranging from .70 to .78 for “agency” and “communion” (see Table 1).

Table 1 Interrater agreement for human coding (n = 1,000).

Regarding label distribution, the human annotators classified most sentences as “none,” with 94% for strong cues and 89% for combined weak + strong cues. Among sentences containing personality cues, “communion” appeared more frequently (approximately 56% of strong cues and 63% of weak + strong cues) than “agency” (around 44% of strong cues and 38% of weak + strong cues).Footnote 7

4.2 Comparison of Computational Scores with Human-annotated Test Set (External Validation)

We next compared how the different computational text analysis models would compare to the human-annotated test set, using macro-F1 score as the performance measure. Figure 3 displays the performance of all models across balanced and unbalanced test and training datasets.Footnote 8 Red points indicate results on the strong datasets that only contain strong personality cues, and blue triangles indicate results on the combined weak + strong datasets with an equal proportion of weak and strong personality cues. Overall, our results show that DeepSeek-V3 and GPT-4o demonstrated the highest performance across dataset variants. As Table 2 shows, DeepSeek-V3 performed best on both balanced and unbalanced datasets, attaining a macro-F1 score of 0.77. GPT-4o performed similarly well, achieving a macro-F1 score of 0.76 on the strong-balanced dataset, and slightly worse performance on the balanced and unbalanced dataset. In contrast, the Llama 3-8B model showed significantly lower performance, with the highest F1 scores reaching approximately 0.60 for the balanced dataset and 0.55 for the unbalanced dataset, along with greater variability across training runs.

Figure 3 Comparison of classifier performance for balanced and unbalanced dataset (n = 600). The y-axis depicts the mean macro-F1 score across three random seeds. Error bars indicate standard deviation across random seeds.

Table 2 Overall best-performing models on the human-annotated test data across three random seeds.

Likewise, models using a semisupervised (XLM-RoBERTa) or supervised (SVM) strategy exhibited also lower macro-F1 scores. The SVM model performed reasonably well on the balanced dataset with strong labels (F1 = 0.63, but with great variability as indicated by the error bars), but performed worse on the weak + strong dataset (F1 = 0.59) and both unbalanced datasets (F1 = 0.55 and 0.61, respectively). The XLM-RoBERTa model showed only slight improvements over SVM for the balanced dataset, with its best performance recorded for the balanced strong-labeled dataset (F1 = 0.70).

4.3 Functional Tests (Structural Validation)

In addition to the human-label comparison, we conduct systematic tests of the different models’ performance on survey items drawn from Lechner et al.’s (Reference Lechner, Aichholzer, Birkenmaier and Bluemke2025) original personality inventory as well as on purposefully chosen sample sentences from the corpus to evaluate whether models can reliably predict explicitly defined test cases (units). Specifically, each model was tested with 18 test units: (a) the three self-report items designed to reflect each associated trait (for instance, “I show understanding for the needs of disadvantaged people” for communion or “I am a born leader” for agency); (b) the three opposing items per trait, expected to be classified as “other” (e.g., “I am hard, cold-hearted” as the opposite pole for communion or “I prefer to stay in the background” for the opposite pole for agency); and (c) six sentences from our original corpus, which we identified as unrelated to any personality cues and that also should be labeled “other” as well (for a full list of test units, see Appendix J of the Supplementary Material).

Our results are displayed in Figure 4. A well-performing model’s confusion matrix should ideally display three green predictions in the lower left corner (representing three correctly predicted “agency” items), three green predictions in the center (corresponding to three correctly predicted “communion” items), and twelve green predictions in the upper right corner (covering the six “other” sentences from the original corpus and the three opposing poles for both “agency” and “communion”, totaling twelve “other” predictions). For strong labels, only few-shot prompting with GPT-4o and DeepSeek-V3 correctly classified all test units, whereas zero-shot prompting for both models was still satisfactory with only minor prediction errors. For combined weak + strong labels, encompassing both weak and strong cues together, only DeepSeek-V3 was able to correctly classify all test units. For zero-shot prompting, the performance of GPT-4o and DeepSeek-V3 was also still satisfactory, again with only minor prediction errors.

Figure 4 Confusion matrices for the test units (n = 18). The y-axis corresponds to the true values, and the x-axis corresponds to the predicted values, with o = “other,” c = “communion,” and a = “agency.”

In contrast, SVM, XML-RoBERTa, and Llama 3-8B showed more overall wrong predictions for both strong and combined weak + strong labels, often failing to reach a passing rate of 50% of all units. Overall, we conclude that validation evidence for accurate and consistent classification was demonstrated exclusively by GPT-4o and DeepSeek-V3.

4.4 Exploring Sensitivity to Partisan Differences (External Validation)

Instead of focusing solely on convergence with human annotations, external validation can also assess whether a measure aligns with expected real-world relationships. We therefore conduct additional validation tests to examine the sensitivity of our measures to partisan variation in the data. Personality and ideology are known to intersect, and personality and value orientations show political correlations that are not always easily separable (Fatke Reference Fatke2017).Footnote 9

To test whether our measures are sensitive to partisan differences in our data, we examine partisan differences in public personality across two use cases: (1) aggregated personality scores of politicians across political parties in the German Bundestag for the full unlabeled corpus and (2) individual personality scores for the U.S. Presidential Debate between Donald Trump and Kamala Harris on September 10, 2024.Footnote 10 In both cases, we expect, in line with previous research, that liberal or left-leaning politicians generally display higher and more universal levels of communion compared to their conservative counterparts, who often tend to limit empathy to deliberatively chosen (in-)groups (Hasson et al. Reference Hasson, Tamir, Brahms, Cohrs and Halperin2018; McCue and Gopoian Reference McCue and Gopoian2000).

For clarity, we report results exclusively for the best-performing model that successfully passed all previous validation steps, namely DeepSeek-V3.

4.4.1 Left-Right Communion Distribution in the Bundestag

Figure 5 presents scatterplots showing the relationship between politicians’ shares of communion with mean politicians’ values aggregated at the party level (y-axis) and their parties’ expert-rated positions on the Chapel Hill Expert Survey (CHES, see Jolly et al. (Reference Jolly2022)) for two dimensions that make up the left–right continuum: the economic left–right (left panel) and the cultural libertarian–authoritarian (right panel) dimension. Generally, the highest shares of sentences containing communion cues can be found for the left party “LINKE” (19%), followed by the Greens (12%) and social–democrat SPD (10%), whereas politicians from more conservative parties like the CDU (8%) and right-wing AFD and liberal FDP (both around 6%) show overall lower communion values.

Figure 5 Correspondence of aggregated politicians’ shares for each party for “communion” (y-axis) and the respective CHES dimension (x-axis). Error bars indicate standard errors around each party mean.

As indicated by the regression lines, we see a clear and consistent negative relationship between the parties CHES score and the politicians’ share of communion. This association is particularly pronounced not only for the economic dimension (left panel), but also present in the cultural dimension (right panel). This pattern suggests that the signaling of communion is systematically more prevalent among politicians from economically and culturally left-leaning parties, in line with our prior assumption. These results support the validity of our agency–communion measures because they vividly demonstrate their correspondence with established ideological dimensions.Footnote 11 In Appendices M and N of the Supplementary Material, we further report descriptive analysis for the data at both the aggregated and individual levels, suggesting numerous avenues for inquiry that more in-depth analyses could explore.Footnote 12

4.4.2 U.S. Presidential Debate Corpus

Next, we measure the signaling of personality traits for the 2024 U.S. Presidential Debate between Donald Trump and Kamala Harris on September 10, 2024, to demonstrate that our measure is also sensitive to individual-level human variation. Transcripts for the debate were derived from the American Presidency Project (Woolley and Peters Reference Woolley and Peters1999), translated from English to German using the Deepl-API and subsequently processed using the same steps as in our previous analysis.

Our findings, visualized in Figure 6, confirm our expectations that Trump should signal more agency compared to Kamala Harris. On average, Trump’s sentences are notably shorter, leading to nearly twice as many utterances (n = 655) compared to Harris (n = 320). Notably, the overall share of personality cues is higher for Harris (approximately 23%) than for Trump (approximately 10%). However, examining the within-person distribution reveals a clear pattern: Trump displays a stronger relative emphasis on agency (63% agency vs. 37% communion), whereas Harris shows the reverse pattern, with a higher relative share of communion (59% vs. 41% agency). This means that the relative prominence of agency and communion is nearly reversed between the two candidates. These results show that our agency–communion measure was able to capture differences in the two candidates’ public personalities that were widely discussed in the media during the election campaign (Hyatt et al. Reference Hyatt, Campbell, Lynam and Miller2018).

Figure 6 Ratio of “agency” and “communion” for the presidential debate on September 10, 2024, between Donald Trump and Kamala Harris.

5 Summary and Conclusion

This study explored the feasibility of using computational text analysis to measure politicians’ publicly displayed personality traits from various textual sources (interviews, speeches, and social media posts) for two traits: agency and communion. Measuring personality traits from verbal behavior is highly challenging, not only because personality traits are inherently fuzzy constructs, making it difficult to identify valid linguistic cues, but also because their linguistic cues are often ambiguous and context dependent, making it difficult for computational models (as well as human raters) to correctly use the right cues to form accurate personality judgments. To address these challenges, we started from a validated personality framework to identify linguistic personality cues. We then compared multiple measurement strategies and models, varying dataset compositions (balanced vs. unbalanced) and personality cue conceptualizations (strong cues only vs. a combination of weak + strong cues). We conducted extensive validation using a systematic framework as outlined in Birkenmaier et al. (Reference Birkenmaier, Wagner and Lechner2024). Our findings suggest that prompt-based methods, particularly GPT-4o and DeepSeek-V3, allowed the most valid measurement of agency and communion from politicians’ public utterances.

Regarding the conceptualization of personality traits, models tasked to detect only strong cues for the traits in question performed better, as expected. This aligns with intuition, as detecting personality cues becomes more challenging when weak cues are included. The fact that personality traits are fuzzy and concepts with ambiguous verbal cues was reflected in the moderate agreement among human coders (between .70 and .78 Krippendorff’s alpha) and the fact that none of the computational models achieved a macro-F1 score above .80 (see Table 1).

Nevertheless, the results suggest that the best available computational methods can indeed capture meaningful personality-related cues in text. These findings provide the foundation for more substantive research in the field of political psychology. For instance, this could include the investigation of changes in politicians’ expressed personality across environmental factors that can indicate how political figures adapt their public personas in response to evolving circumstances. Potential explanatory factors include crisis events, such as economic downturns or global conflicts, which may lead politicians to fulfil the demand for dominant political leaders (Petersen and Laustsen Reference Petersen and Laustsen2020; Sprong et al. Reference Sprong2019). Additionally, individual role transitions—such as taking over the position of party secretary or becoming a lead candidate—could influence the way personality traits are displayed.

While this study demonstrates the potential of computational methods for inferring politicians’ public personality from verbal cues, several limitations remain. The analysis focused on only two personality traits—agency and communion—which should be expanded to other relevant traits. Next, keep in mind that personality cues can be expressed in multiple ways—such as paraverbal and nonverbal features like rhetoric, gestures, or interactions. Thus, future research should complement our approach by using multimodal approaches, incorporating video, audio, and behavioral data (Achmann-Denkler et al. Reference Achmann-Denkler, Fehle, Haim and Wolff2024). Connected to this, our study operated at the sentence level; extending it to paragraph-level or token-level analysis would provide deeper insights into personality expression in political speech. Finally, although our validation did not reveal systematic bias, we acknowledge that large language models can generally reflect biases from their pretraining data. In the case of DeepSeek-V3, recent reports have pointed to potential pro-Chinese bias stemming from data selection or censorship (Huang et al. Reference Huang, Lin, Imbot, Fu and Tu2025). While we found no indication of such bias in our study, we recommend additional validation in studies involving more politically sensitive content or contexts.

In conclusion, our study demonstrates that while measuring personality trait expressions from text remains a daunting task, it is not an impossible one. By combining robust validation strategies with recent advances in language models, we demonstrated that even abstract traits like agency and communion leave detectable traces in political discourse if the right operationalizations are chosen. This opens the door to a new generation of research: one where personality is not just surveyed or inferred, but measured at scale, over time, and in context—from speeches to debates, from parliaments to press conferences. If personality is indeed what politicians project to the public, then computational methods may soon become essential tools in decoding the psychological choreography of power.

Acknowledgments

We thank all those who contributed to this manuscript, especially the participants of the 6th ANNUAL COMPTEXT Conference 2024 in Amsterdam, as well as Brandon Stewart, Co-Editor-in-Chief, and the four outstanding anonymous reviewers, whose efforts went above and beyond to improve the manuscript.

Funding Statement

This research received no particular funding.

Data Availability Statement

Full replication materials for this article are available at: Birkenmaier, Lukas, and Lechner, Clemens. Reference Birkenmaier and Lechner2025, “Replication Data for: Measuring Politicians’ Public Personality Traits using Computational Text Analysis: A Multi-Method Feasibility Study for Agency and Communion”, https://doi.org/10.7910/DVN/MSY9L1, Harvard Dataverse.

Author Contributions

Both authors contributed equally to this work.

Competing Interests

The authors declare none.

Ethical Standards

No concerns regarding ethical standards are reported.

Supplementary material

To view supplementary material for this article, please visit http://doi.org/10.1017/pan.2025.10025.

Footnotes

Edited by: Brandon M. Stewart and Daniel J. Hopkins

1 Full replication materials for this article are available at: Birkenmaier, Lukas, and Lechner, Clemens. Reference Gerber, Huber, Doherty, Dowling and Ha2025, “Replication Data for: Measuring Politicians’ Public Personality Traits Using Computational Text Analysis: A Multi-Method Feasibility Study for Agency and Communion”, https://doi.org/10.7910/DVN/MSY9L1, Harvard Dataverse.

2 This branch of research is based on the observation that personality traits are often linked to ideology (Gerber et al. Reference Gerber, Huber, Doherty, Dowling and Ha2010). For instance, traits like conscientiousness are often linked with conservative ideologies, whereas openness tends to correlate with liberal views (Carney et al. Reference Carney, Jost, Gosling and Potter2008).

3 Although these surveys are designed to provide a direct assessment of politician’s intrinsic personality, in the political domain, they are highly susceptible to social desirability bias and “faking good,” meaning politicians are likely to portray themselves in a more favorable light unintentionally or intentionally. Consequently, we argue that politicians’ self-reports can also be seen as a special type of self-statement personality cues rather than as accurate reflections of intrinsic traits.

4 Although our analysis centers on two particular traits discussed further next, this workflow can be scaled to accommodate any traits. Appendix B of the Supplementary Material presents a more abstract workflow of our approach.

5 For readability purposes, we subsequently refer to them to “agency” and “communion” instead of the prototypical traits “Empathy & Compassion” and “Assertiveness & Will to Take Control.”

6 We focus on the positive poles of the prototypical traits introduced earlier, but not the opposing poles “Passivity & Submissiveness” (-communion) and “Apathy & Indifference” (-agency). This reflects our assumption that politicians usually aim to project personality cues that reinforce a strong and favorable public image.

7 Of the 1,000 coded sentences, one sentence (“One important aspect is that we take action in the markets to ensure that gas and electricity remain affordable for everyone”) was coded as a weak cue for both “agency” and “communion.” Since this overlap occurred in only 1 of 1,000 cases, we subsequently apply multiclass classification for practical reasons, requiring the classifier to assign the most likely category (agency, communion, or none) to each sentence, even if a sentence may still contain cues for more than one category.

8 As outlined in Section 2, only supervised (SVM)/semisupervised (XLM-RoBERTa) models relied on training/fine-tuning data, whereas prompting only included a selection of examples in the prompt (few-shot prompting).

9 An important part of demonstrating convergent validity is ensuring that personality is measured as a construct distinct from ideology. In Appendix K of the Supplementary Material, we show that though our personality codebook shares some features with the Manifesto Project Codebook, which measures ideology, it also contains distinctive elements that capture cues specific to personality traits, even when some of these cues also carry ideological meanings.

10 In the U.S. context, there is even consistent politician-specific evidence suggesting that Donald Trump is regularly rated lower on empathy-related dimensions reflecting communion and higher on conflict-related dimensions representing agency (Hyatt et al. Reference Hyatt, Campbell, Lynam and Miller2018; Nai, Maier, and Vranić (Reference Wright and Tomlinson2021); Wright and Tomlinson Reference Wright and Tomlinson2018).

11 Although not part of our initial analysis due to less apparent theoretical connections in a multiparty parliamentary system, Appendix L of the Supplementary Material presents the aggregated proportions of “agency” and the CHES dimensions, revealing only a weak positive relationship with the economic dimension and no relationship with the cultural dimension.

12 Appendix M of the Supplementary Material outlines the distribution of agency and communion across data source, showing that levels of communion are significantly higher on social media than in interviews, whereas levels of agency show less variation. Appendix N of the Supplementary Material presents the distribution of individual personality scores across differentiated by gender, indicating that female politicians tend to exhibit higher levels of communion, while the distribution of agency appears more heterogeneous across genders.

References

Achmann-Denkler, Michael, Fehle, Jakob, Haim, Mario, and Wolff, Christian. 2024. “Detecting Calls to Action in Multimodal Content: Analysis of the 2021 German Federal Election Campaign on Instagram.” arXiv. https://doi.org/10.48550/ARXIV.2409.02690.CrossRefGoogle Scholar
Aichholzer, Julian, and Willmann, Johanna. 2020. “Desired Personality Traits in Politicians: Similar to Me but More of a Leader.” Journal of Research in Personality 88: 103990.10.1016/j.jrp.2020.103990CrossRefGoogle Scholar
Amsalem, Eran, Fogel-Dror, Yair, Shenhav, Shaul R., and Sheafer, Tamir. 2020. “Fine-Grained Analysis of Diversity Levels in the News.” Communication Methods and Measures 14 (4): 266284. https://doi.org/10.1080/19312458.2020.1825659.CrossRefGoogle Scholar
Azucar, Danny, Marengo, D., and Settanni, M.. 2018. “Predicting the Big 5 Personality Traits from Digital Footprints on Social Media: A Meta-Analysis.” Personality and Individual Differences 124: 150159. https://doi.org/10.1016/J.PAID.2017.12.018.CrossRefGoogle Scholar
Bakan, David. 1966. “The Duality of Human Existence: An Essay on Psychology and Religion.” Rand McNally. https://psycnet.apa.org/record/1967-02200-000.Google Scholar
Best, Heinrich. 2011. “Does Personality Matter in Politics? Personality Factors as Determinants of Parliamentary Recruitment and Policy Preferences.” Comparative Sociology 10 (6): 928948.10.1163/156913311X607638CrossRefGoogle Scholar
Birkenmaier, Lukas, Wagner, Claudia, and Lechner, Clemens. 2024. “ValiText: A Unified Validation Framework for Computational Text-Based Measures of Social Constructs.” arXiv. https://doi.org/10.48550/arXiv.2307.02863.CrossRefGoogle Scholar
Birkenmaier, Lukas, and Lechner, Clemens. 2025, “Replication Data for: Measuring Politicians’ Public Personality Traits Using Computational Text Analysis: A Multi-Method Feasibility Study for Agency and Communion", https://doi.org/10.7910/DVN/MSY9L1, Harvard Dataverse.CrossRefGoogle Scholar
Birkenmaier, Lukas, Laureen, Sieber, and Felix, Bergstein. 2025. “PolInterviews: A dataset of German politician public broadcast interviews” arXiv. https://arxiv.org/abs/2501.04484.Google Scholar
Birkenmaier, Lukas, Wurthmann, Constantin, and Sältzer, Marius. 2025. “Between Home Turf and Hinterland: Directly Elected MPs Focus More on Local and Deprived Places than List Candidates on Social Media.” Working Paper.Google Scholar
Bittner, Amanda, and Peterson, David AM. 2018. “Introduction: Personality, Party Leaders, and Election Campaigns.” Electoral Studies 54: 237239.10.1016/j.electstud.2018.04.005CrossRefGoogle Scholar
Brunswik, Egon. 1952. The Conceptual Framework of Psychology. Chicago, IL: University of Chicago Press.Google Scholar
Caprara, Gian Vittorio, Barbaranelli, Claudio, Consiglio, Chiara, Picconi, Laura, and Zimbardo, Philip G.. 2003. “Personalities of Politicians and Voters: Unique and Synergistic Relationships.” Journal of Personality and Social Psychology 84 (4): 849. https://doi.org/10.1037/0022-3514.84.4.849.CrossRefGoogle ScholarPubMed
Caprara, Gian Vittorio, and Zimbardo, Philip G.. 2004. “Personalizing Politics: A Congruency Model of Political Preference.” American Psychologist 59 (7): 581. https://doi.org/10.1037/0003-066X.59.7.581.CrossRefGoogle ScholarPubMed
Carney, Dana R., Jost, John T., Gosling, Samuel D., and Potter, Jeff. 2008. “The Secret Lives of Liberals and Conservatives: Personality Profiles, Interaction Styles, and the Things They Leave Behind.” Political Psychology 29 (6): 807840. https://doi.org/10.1111/j.1467-9221.2008.00668.x.CrossRefGoogle Scholar
Chen, Jiayu, Qiu, Lin, and Ho, Moon-Ho Ringo. 2020. “A Meta-Analysis of Linguistic Markers of Extraversion: Positive Emotion and Social Process Words.” Journal of Research in Personality 89 (December): 104035. https://doi.org/10.1016/j.jrp.2020.104035.CrossRefGoogle Scholar
Colliander, Jonas, Marder, Ben, Falkman, Lena Lid, Madestam, Jenny, Modig, Erik, and Sagfossen, Sofie. 2017. “The Social Media Balancing Act: Testing the Use of a Balanced Self-Presentation Strategy for Politicians Using Twitter.” Computers in Human Behavior 74: 277285.10.1016/j.chb.2017.04.042CrossRefGoogle Scholar
Conneau, Alexis, et al. 2020. “Unsupervised Cross-Lingual Representation Learning at Scale.” In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 84408451. https://doi.org/10.18653/v1/2020.acl-main.747.CrossRefGoogle Scholar
Costa, Paul T. Jr., and McCrae, Robert R.. 1995. “Domains and Facets: Hierarchical Personality Assessment Using the Revised NEO Personality Inventory.” Journal of Personality Assessment 64 (1): 2150. https://doi.org/10.1207/s15327752jpa6401_2.CrossRefGoogle ScholarPubMed
DeepSeek-AI, Aixin Liu, et al. 2024. “DeepSeek-V3 Technical Report.” arXiv. https://doi.org/10.48550/arXiv.2412.19437.CrossRefGoogle Scholar
Entringer, Theresa M., Gebauer, Jochen E., and Paulhus, Delroy L.. 2022. “Extracting Agency and Communion From the Big Five: A Four-Way Competition.” Assessment 29 (6): 12161235. https://doi.org/10.1177/10731911211003978.CrossRefGoogle ScholarPubMed
Fatke, Matthias. 2017. “Personality Traits and Political Ideology: A First Global Assessment.” Political Psychology 38 (5): 881899. https://doi.org/10.1111/pops.12347.CrossRefGoogle Scholar
Federico, Christopher M., and Malka, Ariel. 2018. “The Contingent, Contextual Nature of the Relationship Between Needs for Security and Certainty and Political Preferences: Evidence and Implications.” Political Psychology 39 (S1): 348. https://doi.org/10.1111/pops.12477.CrossRefGoogle Scholar
Fenno, Richard F. 2003. Home Style: House Members in Their Districts. Longman Classics in Political Science. New York: Longman.Google Scholar
Funder, David C. 1995. “On the Accuracy of Personality Judgment: A Realistic Approach.” Psychological Review 102 (4): 652670. https://doi.org/10.1037/0033-295X.102.4.652.CrossRefGoogle ScholarPubMed
Funk, Carolyn L. 1997. “Implications of Political Expertise in Candidate Trait Evaluations.” Political Research Quarterly 50 (3): 675697. https://doi.org/10.1177/106591299705000309.CrossRefGoogle Scholar
George, A.L., and George, J.L. 1998. Presidential Personality and Performance. Boulder, CO: WestviewPress.Google Scholar
Gerber, Alan S., Huber, Gregory A., Doherty, David, Dowling, Conor M., and Ha, Shang E.. 2010. “Personality and Political Attitudes: Relationships across Issue Domains and Political Contexts.” American Political Science Review 104 (1): 111133.10.1017/S0003055410000031CrossRefGoogle Scholar
Gilardi, Fabrizio, Alizadeh, Meysam, and Kubli, Maël. 2023. “ChatGPT Outperforms Crowd Workers for Text-Annotation Tasks.” Proceedings of the National Academy of Sciences 120 (30): e2305016120. https://doi.org/10.1073/pnas.2305016120.CrossRefGoogle ScholarPubMed
Goffman, Erving. 2016. “The Presentation of Self in Everyday Life.” In: Longhofer, Wesley, Winchester, Daniel. Social Theory Re-Wired, 482493. New York: Routledge.Google Scholar
Gulati, Girish J. 2004. “Members of Congress and Presentation of Self on the World Wide Web.” Harvard International Journal of Press/Politics 9 (1): 2240. https://doi.org/10.1177/1081180X03259758.CrossRefGoogle Scholar
Hanania, Richard. 2017. “The Personalities of Politicians: A Big Five Survey of American Legislators.” Personality and Individual Differences 108: 164167.10.1016/j.paid.2016.12.020CrossRefGoogle Scholar
Hasson, Yossi, Tamir, Maya, Brahms, Kea S., Cohrs, J. Christopher, and Halperin, Eran. 2018. “Are Liberals and Conservatives Equally Motivated to Feel Empathy Toward Others?Personality and Social Psychology Bulletin 44 (10): 14491459. https://doi.org/10.1177/0146167218769867.CrossRefGoogle ScholarPubMed
Hayes, Danny. 2009. “Has Television Personalized Voting Behavior?Political Behavior 31 (2): 231260. https://doi.org/10.1007/s11109-008-9070-0.CrossRefGoogle Scholar
Hermann, Margaret G. 1980. “Explaining Foreign Policy Behavior Using the Personal Characteristics of Political Leaders.” International Studies Quarterly 24 (1): 746. https://doi.org/10.2307/2600126.CrossRefGoogle Scholar
Hogan, Robert. 1982. “A Socioanalytic Theory of Personality.” In M. M. Page. Nebraska Symposium on Motivation, 5589. Lincoln, NE: University of Nebraska Press.Google Scholar
Huang, PeiHsuan, Lin, ZihWei, Imbot, Simon, Fu, WenCheng, and Tu, Ethan. 2025. “Analysis of LLM Bias (Chinese Propaganda & Anti-US Sentiment) in DeepSeek-R1 vs. ChatGPT O3-Mini-High.” arXiv. https://doi.org/10.48550/arXiv.2506.01814.CrossRefGoogle Scholar
Hyatt, Courtland, Campbell, W. Keith, Lynam, Donald R., and Miller, Joshua D.. 2018. “Dr. Jekyll or Mr. Hyde? President Donald Trump’s Personality Profile as Perceived from Different Political Viewpoints.” Collabra: Psychology 4 (1): 29.10.1525/collabra.162CrossRefGoogle Scholar
Jolly, Seth, et al. 2022. “Chapel Hill Expert Survey Trend File, 1999–2019.” Electoral Studies 75 (February): 102420. https://doi.org/10.1016/j.electstud.2021.102420.CrossRefGoogle Scholar
Joly, Jeroen, Soroka, Stuart, and Loewen, Peter. 2019. “Nice Guys Finish Last: Personality and Political Success.” Acta Politica 54 (4): 667683. https://doi.org/10.1057/s41269-018-0095-z.CrossRefGoogle Scholar
Krippendorff, Klaus. 2018. Content Analysis: An Introduction to Its Methodology. Sage Publications. Thousand Oaks, California, USA.Google Scholar
Laurer, Moritz, van Atteveldt, Wouter, Casas, Andreu, and Welbers, Kasper. 2024. “Less Annotating, More Classifying: Addressing the Data Scarcity Issue of Supervised Machine Learning with Deep Transfer Learning and BERT-NLI.” Political Analysis 32 (1): 84100. https://doi.org/10.1017/pan.2023.20.CrossRefGoogle Scholar
Laustsen, Lasse, and Bor, Alexander. 2017. “The Relative Weight of Character Traits in Political Candidate Evaluations: Warmth Is More Important than Competence, Leadership and Integrity.” Electoral Studies 49: 96107.10.1016/j.electstud.2017.08.001CrossRefGoogle Scholar
Lechner, Clemens, Aichholzer, Julian, Birkenmaier, Lukas, and Bluemke, Matthias. 2025. “Mapping the Political Personality (PoliPers): An Integrative Framework for Assessing Ideal and Actual Personality Profiles in Politicians.” OSF. https://doi.org/10.31234/osf.io/w42pj.CrossRefGoogle Scholar
Lee, Jayeon, and Lim, Young-shin. 2016. “Gendered Campaign Tweets: The Cases of Hillary Clinton and Donald Trump.” Public Relations Review 42 (5): 849855. https://doi.org/10.1016/j.pubrev.2016.07.004.CrossRefGoogle Scholar
Lilienfeld, Scott O., Waldman, Irwin D., Landfield, Kristin, Watts, Ashley L., Rubenzer, Steven, and Faschingbauer, Thomas R.. 2012. “Fearless Dominance and the US Presidency: Implications of Psychopathic Personality Traits for Successful and Unsuccessful Political Leadership.” Journal of Personality and Social Psychology 103 (3): 489.10.1037/a0029392CrossRefGoogle ScholarPubMed
Mair, Peter. 2025. “Democracy beyond parties.” CSD Working Papers. https://escholarship.org/uc/item/3vs886v9.Google Scholar
Maier, Jürgen, Oschatz, Corinna, Stier, Sebastian, and Zettler, Ingo. 2023. “A Short Scale to Measure Self-Reported Aversive Personality in Political Elites.” Journal of Personality Assessment 105 (5): 625635. https://doi.org/10.1080/00223891.2022.2130341.CrossRefGoogle Scholar
Mairesse, François, Walker, Marilyn A., Mehl, Matthias R., and Moore, Roger K.. 2007. “Using Linguistic Cues for the Automatic Recognition of Personality in Conversation and Text.” Journal of Artificial Intelligence Research 30: 457500.10.1613/jair.2349CrossRefGoogle Scholar
McAllister, Ian. 2007. “The Personalization of Politics.” In The Oxford Handbook of Political Behavior, edited by Dalton, Russell J. and Klingemann, Hans-Dieter, 571588. Oxford: Oxford University Press. https://doi.org/10.1093/oxfordhb/9780199270125.003.0030.Google Scholar
McCue, Clifford P., and Gopoian, J. David. 2000. “Dispositional Empathy and the Political Gender Gap.” Women & Politics 21 (2): 120. https://doi.org/10.1300/J014v21n02_01.CrossRefGoogle Scholar
McDonald, Jennifer Dodorico. 2008. “Measuring Personality Constructs: The Advantages and Disadvantages of Self-Reports, Informant Reports and Behavioural Assessments.” Enquire 1 (1): 119.Google Scholar
Miller, Arthur H., Wattenberg, Martin P., and Malanchuk, Oksana. 1986. “Schematic Assessments of Presidential Candidates.” American Political Science Review 80 (2): 521540.10.2307/1958272CrossRefGoogle Scholar
Mondak, Jeffery J. 1995. “Competence, Integrity, and the Electoral Success of Congressional Incumbents.” The Journal of Politics 57 (4): 10431069. https://doi.org/10.2307/2960401.CrossRefGoogle Scholar
Nai, Alessandro. 2019. “The Electoral Success of Angels and Demons: Big Five, Dark Triad, and Performance at the Ballot Box.” Journal of Social and Political Psychology 7 (2): 830862.10.5964/jspp.v7i2.918CrossRefGoogle Scholar
Nai, Alessandro, and Maier, Jürgen. 2024. Dark Politics: The Personality of Politicians and the Future of Democracy. New York, NY: Oxford University Press.10.1093/oso/9780197681756.001.0001CrossRefGoogle Scholar
Nai, A., Maier, J., and Vranić, J. 2021. “Personality Goes a Long Way (for Some). An Experimental Investigation Into Candidate Personality Traits, Voters’ Profile, and Perceived Likeability.” Frontiers in Political Science, 3 https://doi.org/10.3389/fpos.2021.636745.CrossRefGoogle Scholar
Nørgaard, Asbjørn Sonne, and Klemmensen, Robert. 2019. “The Personalities of Danish MPs: Trait- and Aspect-Level Differences.” Journal of Personality 87 (2): 267275. https://doi.org/10.1111/jopy.12388.CrossRefGoogle ScholarPubMed
Park, Gregory, et al. 2015. “Automatic Personality Assessment through Social Media Language.” Journal of Personality and Social Psychology 108 (6): 934952. https://doi.org/10.1037/pspp0000020.CrossRefGoogle ScholarPubMed
Partsch, Melanie V, Bluemke, Matthias, and Lechner, Clemens M. 2022. “Revisiting the Hierarchical Structure of the 24 VIA Character Strengths: Three Global Dimensions May Suffice to Capture Their Essence.” European Journal of Personality 36 (5): 825845. https://doi.org/10.1177/08902070211017760.CrossRefGoogle Scholar
Pedersen, Helene Helboe, and Rahat, Gideon. 2021. “Political Personalization and Personalized Politics within and beyond the Behavioural Arena.” Party Politics 27 (2): 211219. https://doi.org/10.1177/1354068819855712.CrossRefGoogle Scholar
Perloff, Richard M. 2013. Political Communication: Politics, Press, and Public in America. New York: Routledge. https://doi.org/10.4324/9780203811832.CrossRefGoogle Scholar
Petersen, Michael Bang, and Laustsen, Lasse. 2020. “Dominant Leaders and the Political Psychology of Followership.” Current Opinion in Psychology 33: 136141.10.1016/j.copsyc.2019.07.005CrossRefGoogle ScholarPubMed
Pillai, Rajnandini, Williams, Ethlyn A, Lowe, Kevin B, and Jung, Dong I. 2003. “Personality, Transformational Leadership, Trust, and the 2000 U.S. Presidential Vote.” The Leadership Quarterly 14 (2): 161192. https://doi.org/10.1016/S1048-9843(03)00008-0.CrossRefGoogle Scholar
Rahat, Gideon, and Kenig, Ofer. 2018. From Party Politics to Personalized Politics?: Party Change and Political Personalization in Democracies. Oxford: Oxford University Press.10.1093/oso/9780198808008.001.0001CrossRefGoogle Scholar
Ramey, Adam J., Klingler, Jonathan D., and Hollibaugh, Gary E.. 2019. “Measuring Elite Personality Using Speech.” Political Science Research and Methods 7 (1): 163184.10.1017/psrm.2016.12CrossRefGoogle Scholar
Rauh, Christian, and Schwalbach, Jan. 2020. “The ParlSpeech V2 Data Set: Full-Text Corpora of 6.3 Million Parliamentary Speeches in the Key Legislative Chambers of Nine Representative Democracies.” Harvard Dataverse. https://doi.org/10.7910/DVN/L4OAKN.CrossRefGoogle Scholar
Rubenzer, Steven J., Faschingbauer, Thomas R., and Ones, Deniz S.. 2000. “Assessing the U.S. Presidents Using the Revised NEO Personality Inventory.” Assessment 7 (4): 403419. https://doi.org/10.1177/107319110000700408.CrossRefGoogle ScholarPubMed
Saucier, Gerard, et al. 2014. “A Basic Bivariate Structure of Personality Attributes Evident Across Nine Languages.” Journal of Personality 82 (1): 114. https://doi.org/10.1111/jopy.12028.CrossRefGoogle ScholarPubMed
Schoen, Harald, and Schumann, Siegfried. 2005. “Missing Links? Zur Position von Persönlichkeitsfaktoren in Ansätzen zur Erklärung von Wahlverhalten.” In Wahlen und Wähler, edited by Falter, Jürgen W., Gabriel, Oscar W., and Weßels, Bernhard, 388413. Wiesbaden: VS Verlag für Sozialwissenschaften. https://doi.org/10.1007/978-3-322-80516-4_16.CrossRefGoogle Scholar
Schoonvelde, Martijn, Schumacher, Gijs, and Bakker, Bert N.. 2019. “Friends With Text as Data Benefits: Assessing and Extending the Use of Automated Text Analysis in Political Science and Political Psychology.” Journal of Social and Political Psychology 7 (1): 124143. https://doi.org/10.5964/jspp.v7i1.964.CrossRefGoogle Scholar
Schumacher, Gijs, and Zettler, Ingo. 2019. “House of Cards or West Wing? Self-Reported HEXACO Traits of Danish Politicians.” Personality and Individual Differences 141 (April): 173181. https://doi.org/10.1016/j.paid.2018.12.028.CrossRefGoogle Scholar
Schürmann, Lennart, and Stier, Sebastian. 2022. “Who Represents the Constituency? Online Political Communication by Members of Parliament in the German Mixed-Member Electoral System.” Legislative Studies Quarterly, March, lsq.12379. https://doi.org/10.1111/lsq.12379.Google Scholar
Scott, Colin, and Medeiros, Mike. 2020. “Personality and Political Careers: What Personality Types Are Likely to Run for Office and Get Elected?Personality and Individual Differences 152 (January): 109600. https://doi.org/10.1016/j.paid.2019.109600.CrossRefGoogle Scholar
Sprong, Stefanie, et al. 2019. “‘Our Country Needs a Strong Leader Right Now’: Economic Inequality Enhances the Wish for a Strong Leader.” Psychological Science 30 (11): 16251637. https://doi.org/10.1177/0956797619875472.CrossRefGoogle Scholar
Stanyer, James. 2008. “Elected Representatives, Online Self-Presentation and the Personal Vote: Party, Personality and Webstyles in the United States and United Kingdom.” Information, Communication & Society 11 (3): 414432. https://doi.org/10.1080/13691180802025681.CrossRefGoogle Scholar
Stier, Sebastian, Bleier, Arnim, Lietz, Haiko, and Strohmaier, Markus. 2018. “Election Campaigning on Social Media: Politicians, Audiences, and the Mediation of Political Communication on Facebook and Twitter.” Political Communication 35 (1): 5074. https://doi.org/10.1080/10584609.2017.1334728.CrossRefGoogle Scholar
Thielmann, Isabel, and Hilbig, Benjamin E.. 2019. “Nomological Consistency: A Comprehensive Test of the Equivalence of Different Trait Indicators for the Same Constructs.” Journal of Personality 87 (3): 715730. https://doi.org/10.1111/jopy.12428.CrossRefGoogle ScholarPubMed
Touvron, Hugo, et al. 2023. “LLaMA: Open and Efficient Foundation Language Models.” arXiv. https://doi.org/10.48550/arXiv.2302.13971.CrossRefGoogle Scholar
Van Aelst, Peter, Sheafer, Tamir, and Stanyer, James. 2012. “The Personalization of Mediated Political Communication: A Review of Concepts, Operationalizations and Key Findings.” Journalism 13 (2): 203220. :https://doi.org/10.1177/1464884911427802.CrossRefGoogle Scholar
Van Zoonen, Liesbet. 2005. Entertaining the Citizen: When Politics and Popular Culture Converge. Lanham, Maryland: Rowman & Littlefield.Google Scholar
Visser, Beth A., Book, Angela S., and Volk, Anthony A.. 2017. “Is Hillary Dishonest and Donald Narcissistic? A HEXACO Analysis of the Presidential Candidates’ Public Personas.” Personality and Individual Differences 106: 281286.10.1016/j.paid.2016.10.053CrossRefGoogle Scholar
Vries, Reinout E. de, and van Prooijen, Jan-Willem. 2019. “Voters Rating Politicians’ Personality: Evaluative Biases and Assumed Similarity on Honesty-Humility and Openness to Experience.” Personality and Individual Differences 144: 100104.10.1016/j.paid.2019.03.001CrossRefGoogle Scholar
Walker, Stephen G., Schafer, Mark, and Young, Michael D.. 1998. “Systematic Procedures for Operational Code Analysis: Measuring and Modeling Jimmy Carter’s Operational Code.” International Studies Quarterly 42 (1): 175189. https://doi.org/10.1111/0020-8833.00074.CrossRefGoogle Scholar
Watts, Ashley L., et al. 2013. “The Double-Edged Sword of Grandiose Narcissism: Implications for Successful and Unsuccessful Leadership Among U.S. Presidents.” Psychological Science 24 (12): 23792389. https://doi.org/10.1177/0956797613491970.CrossRefGoogle Scholar
Widmann, Tobias, and Wich, Maximilian. 2022. “Creating and Comparing Dictionary, Word Embedding, and Transformer-Based Models to Measure Discrete Emotions in German Political Text.” Political Analysis (June): 116. https://doi.org/10.1017/pan.2022.15.CrossRefGoogle Scholar
Winter, David G. 2005. “Things I’ve Learned About Personality From Studying Political Leaders at a Distance.” Journal of Personality 73 (3): 557584. https://doi.org/10.1111/j.1467-6494.2005.00321.x.CrossRefGoogle ScholarPubMed
Winter, David G., Hermann, Margaret G., Weintraub, Walter, and Walker, Stephen G.. 1991. “The Personalities of Bush and Gorbachev Measured at a Distance: Procedures, Portraits, and Policy.” Political Psychology 12 (2): 215245. https://doi.org/10.2307/3791463.CrossRefGoogle Scholar
Woolley, John T., and Peters, Gerhard. 1999. “The American Presidency Project.” Santa Barbara, CA. https://www.jfkthelastspeech.org/wp-content/uploads/2023/11/JFK-Speech-at-San-Diego-State.pdf.Google Scholar
Wright, Joshua D., and Tomlinson, Monica F.. 2018. “Personality Profiles of Hillary Clinton and Donald Trump: Fooled by Your Own Politics.” Personality and Individual Differences 128: 2124.10.1016/j.paid.2018.02.019CrossRefGoogle Scholar
Zettler, Ingo, Thielmann, Isabel, Hilbig, Benjamin E., and Moshagen, Morten. 2020. “The Nomological Net of the HEXACO Model of Personality: A Large-Scale Meta-Analytic Investigation.” Perspectives on Psychological Science 15 (3): 723760. https://doi.org/10.1177/1745691619895036.CrossRefGoogle ScholarPubMed
Ziems, Caleb, Held, William, Shaikh, Omar, Chen, Jiaao, Zhang, Zhehao, and Yang, Diyi. 2024. “Can Large Language Models Transform Computational Social Science?Computational Linguistics 50 (1): 237291.10.1162/coli_a_00502CrossRefGoogle Scholar
Figure 0

Figure 1 Conceptual model (adapted from Brunswik, 1952).

Figure 1

Figure 2 Workflow.

Figure 2

Table 1 Interrater agreement for human coding (n = 1,000).

Figure 3

Figure 3 Comparison of classifier performance for balanced and unbalanced dataset (n = 600). The y-axis depicts the mean macro-F1 score across three random seeds. Error bars indicate standard deviation across random seeds.

Figure 4

Table 2 Overall best-performing models on the human-annotated test data across three random seeds.

Figure 5

Figure 4 Confusion matrices for the test units (n = 18). The y-axis corresponds to the true values, and the x-axis corresponds to the predicted values, with o = “other,” c = “communion,” and a = “agency.”

Figure 6

Figure 5 Correspondence of aggregated politicians’ shares for each party for “communion” (y-axis) and the respective CHES dimension (x-axis). Error bars indicate standard errors around each party mean.

Figure 7

Figure 6 Ratio of “agency” and “communion” for the presidential debate on September 10, 2024, between Donald Trump and Kamala Harris.

Supplementary material: File

Birkenmaier and Lechner supplementary material

Birkenmaier and Lechner supplementary material
Download Birkenmaier and Lechner supplementary material(File)
File 605.5 KB
Supplementary material: Link

Birkenmaier and Lechner Dataset

Link