Framing attributions

A methodological and conceptual contribution to the study of explanations in political text

Paulina García-Corral

PhD dissertation defense

Introduction

Explanations are fundamental part of the social world.

Overarching theories across paradigms consider why attributing that “A causes B” matters in our understanding of the social world (Stone 1989; Shiller 2019; Entman 1993)

Examples

Causal reasoning in politics

Selective attribution helps explain how people assign causes of social issues, and how to solve them (Sahar 2014; Rudolph 2006; McCabe 2016).

In politics, Partisan Motivated Reasoning is the mechanism behind why we tend to assign responsibility for positive outcomes to our own party, while blaming bad outcomes on opponents.

Causal reasoning in politics

However, attribution is not just a psychological process, but also a discursive practice:

Attributions as empirical objects

Graham and Singh (2023)

Attributions as empirical objects

Understanding the mechanisms behind selective attribution is being addressed by improving experimental design (Graham and Singh 2023; Tappin, Pennycook, and Rand 2021).

However, questions of external validity persist.

Attributions as empirical objects

When analyzing attributions as they are made in political discourse, content analysis relies on domain specific coding.

Citation Coding Scheme
Sotirovic (2003) Individual vs. Societal
Rozenas and Stukal (2019) National vs. Foreign
Hameleers, Bos, and Vreese (2017) Blame vs. Claim
Tilley and Hobolt (2011) National vs. Regional
Hameleers, Bos, and Vreese (2017) Us vs. Them

Attributions as empirical objects

These two approaches have limited the understanding of the broader structure of causal attribution.


We need a general, scalable method to answer the question:

“who is claiming what and about whom?”

How to measure attributions?

Text-as-data methods are well-suited to capture discursive objects in natural occurring text data.

Causal Language Modeling (CLM) offers a systematic approach to identifying and analyzing causal claims.

Research question


Methodological:

How can causal attributions be used to empirically study political discourse?


Conceptual:

How are causal attribution employed in media framing?

Challenges

  • Causal language is very hard to model.

  • CLM has not been tested in political text data.

  • Rhetorical, strategic, and often contested, and highly contextual.

  • Consensus about basic causal stories is fragmented.

  • The task is not to determine if statements are true, but to parse as expressed.

Operationalize attributions


Break down causal attributions to its linguistic components And get structured data from unstructured text

Instrumentalize attributions

Causal Language Modeling:

  • Define what causal language looks like in text

    • Annotation codebook

    • Corpus creation

  • Train a model to capture these expressions

    • Classification and Span Detection tasks

PolitiCAUSE

PolitiCAUSE results

Large Language Models

Larger contexts promise better performance.

This raises a memorization v.s. generalization question between linguistic cues and causal mechanisms (Kıcıman et al. 2023; Hobbhahn, Lieberum, and Seiler 2022; Takayanagi et al. 2024).

Large Language Models

Large Language Models Evaluation

Designed an evaluation to determine if models can parse causal expressions: Linguistic Causal Disambiguation.

Used adversarial testing to begin understanding the mechanism used to perform a “causal classification” task when using token generation.

Large Language Models results

Lessons learned: Language Modeling

As a method for political analysis

Causal Language Modeling Framework for the study of attribution in political discourse.

Transformer-based models provide a balance between transferability, scalability, and accuracy with modest training data needs.

Provide a probabilistic output perfect for downstream analysis

Method

  1. BERT (Devlin et al. 2018) for Sequence Classification

  1. BERT for Information Extraction

Probabilistic outcomes

Strengths and opportunities

  • Adaptable for specific constructions.
  • Flexible to model choice.
  • Downstream analysis agnostic.
  • Based on semantics.


  • Annotation is cognitively taxing.
  • Sensitive to preprocessing choices.
  • Computational requirements.

Application: Media frames

Causal attributions are seldom operationalized as frames, despite being central to the theory (Entman 1993).

Using causal attributions as frames, I seek answer the question:

How are causal attribution employed in media framing?

To answer this question, I propose non-attribution as a unique strategy.

Non-attribution

Events that have a known cause and a known outcome such as:

Event 1 \(\rightarrow\) Event 2
Russian strike causes 20 people are not alive anymore

Non-attribution

Structure

Type

Example

cause event \(\rightarrow\) effect event explicit causal 20 killed in Kharkiv strike by Russia.
\(\otimes\) \(\rightarrow\) effect event cause implicit 20 killed in Kharkiv.
event - event outcome implicit 20 dead in strike.

Case study: Corpus and data

  • AJ, BBC and CNN
  • October 2023 to June 2025
Region Source Count (Proportion)
ME AJ 1994 (0.29)
BBC 2781 (0.41)
CNN 2017 (0.30)
EE AJ 3603 (0.39)
BBC 2625 (0.29)
CNN 2962 (0.32)

Causal Language Model ft. on news corpus

Is there a difference in use of causal headlines?

How is causality assigned across actors?

Observable Implications

Regime Citation Finding
State Autocratic Rozenas and Stukal (2019) Selective attribution serves to explain social material issues.
Private Democratic Iyengar (1987) Selective attribution serves as partisan filtering of events
PSM Democratic Neutrality performance under soft political power?

Conclusions

Methodological: How can causal attributions be used to empirically study political discourse?

  • Introduce a semi-supervised, NLP-based framework for identifying and measuring causal attributions in political discourse.

Theoretical: How are causal attribution strategically used in media framing?

  • Non-attribution as a framing strategy

By advancing a methodological question I show how we can extend theory of framing and attribution.

My contributions

  • Create a corpus and annotation scheme to detect causal attributions in political text.

  • Develop an evaluation framework for causal language extraction for LMs.

  • Introduce a supervised, NLP-based framework to identify and measure causal attributions in political texts.

  • Analyze how attributions are used in media and present non-attribution as a framing strategy for neutrality performance.

Thank you



Paulina García-Corral

corral@hertie-school.org



Appendix

Paper 1: Background

Dataset Citation Description
BeCAUSE 2.0 Dunietz (n.d.) 5,380 based on construction grammar, newspaper articles, PDT, and US congress.
Parallel Wikipedia Corpus Hidey and McKeown (2016) 265,627 Wikipedia based on alternative lexicalizations.
Causal-TimeBank Mirza et al. (2014) TempEval-3 task + causal singals and links.
EventStoryLine Corpus Caselli and Vossen (2017) 258 documents temporal and causal language of stories
Causal News Corpus Tan et al. (2022) 3,559 events from protest news.
Unicausal Tan, Zuo, and Ng (2023) 58,720 sentences from different datasets
Chemical Induced Disease Gu, Qian, and Zhou (2016) Relation extraction for drugs and adverse effects
FinCausal Mariko et al. (n.d.) Financial documents for causal extraction
PubMed Corpus Yu, Li, and Wang (n.d.) 3,000 research conclusion sentences for correlational v causal language
e-CARE Du et al. (2022) question answering task dataset

Paper 1: Corpus


Corpus composition

tokens (word-level)
N mean std
UNGD 8,872 2,702.87 1,357.05
UK Press 429 787.78 477.58


Finetuning

Total Not Causal Causal
train 12,446 8,897 3,549
val 2,667 1,906 761
test 2,667 1,906 760

Paper 1: Results

BERT RoBERTa DistilBERT UniCausal
Acc 0.832 0.836 0.832 0.715
Prec 0.671 0.686 0.696 0.500
Recall 0.805 0.783 0.730 0.612
MCC 0.617 0.617 0.594 0.550
F1 0.732 0.731 0.712 0.349

Paper 1: Error Analysis

Full Corpus Error Subset
Label TN TP FN FP
Mean conf 4.63 4.13 4.20 4.30
Maj label 0.79 0.10 0.71 0.26
Mean len 26.13 31.73 30.10 31.35
Causal conn 0.37 0.50 0.44 0.48

Paper 1: Discussion

  • Explicit causal sentences only: Implicit structures are also rhetorical and possibly intentional.

  • Single sentence causal structures: Causal attributions also exist across sentences.

  • Corpus external validity from UN debates: Elite rhetoric, audience is other elites, prepared.

Paper 2: Background

Citation Model Findings
Takayanagi et al. (2024) ChatGPT baseline proficency, but can be outperformed by smaller LM
Hobbhahn, Lieberum, and Seiler (2022) GPT-3 Importance of prompting
Gao et al. (2023) ChatGPT Can provide causal explanations but reasoning is unclear
Kıcıman et al. (2023) GPT-3 Outperforms in causal reasoning tasks
Ho et al. (2022) GPT-3 Low rating for cause-effect pair matching
Jin et al. (2024) Llama, Alpaca, GPT High performance for causal stories

Paper 2: Corpus and data

Data N Text
PolitiCAUSE 527 mostly UNGD
Fake news 50 multiple fake news detection databases
Post-train date 50 LexisNexis news databases

Paper 2: Experimental design

Prompt

System: You are a causal language model that performs causal sequence classification and causal span detection. You will classify a text as causal or not causal, and if it’s causal you will extract the causes and effects. The output should be a json with label 1 or 0, cause, and effect value such as {”label”:, ”cause”: ,”effect”: }.

User: But to pay for it, we had to take on debt, precipitated by massive reduction in Government revenue.

Assistant:{”label”: 1,”cause”: ”massive reduction in Government revenue”,
n ”effect”: ”had to take on debt”}

Models

Model parameters

Parameter Value
Temperature 0.0
Top p: 1.0
Top k: 1.0
Frequency penalty 0.0
Presence pentaly 0.0
Repetition penalty 0.0
Max body tokens 200

Inference

  • All Open AI models were run using the Open AI Batch API.

  • Llama and Gemma models were accessed via the Transformers library from Hugging Face, and inference was run using the Together AI API.

  • Gemini-1.5 was run using Google’s AI Studio API.

  • For Google models, the HarmBlockThreshold in the safety settings parameter was set to None for the first two experiments, and set to default for the Fake news and post-training events set of sentences in LCD evaluation.

Paper 2: Results

PolitiCAUSE Causal Sequence Classification results.

LCD Causal Sequence Classification results for Fake News and Post-training events sentence sets.

Paper 2: Error Analysis

Model’s with more parameters exhibit better performance overall:

  • Following recall and precision smaller models may be relying on more explicit causal markers, while larger models can “infer from context”.

  • Scale contributes not only to broader generalization, but also to more faithful alignment with syntactic and semantic cues in discourse.

  • Smaller models appear more susceptible to heuristic pattern matching and are prone to overgeneralizing causal signals from pretraining data.

Paper 2: Discussion

  • PolitiCAUSE contains extensive but noisy data, so only the highest-quality, human-verified sentences were used to ensure reliable evaluation and avoid cases where human annotations might be wrong.

  • The prompt follows established structures from prior work to maximize stability in zero-shot settings, acknowledging that alternative prompting strategies (e.g., Chain of Thought) could also work.

  • Although paraphrasing can also show correct causal structure identification, the experiment required outputs in a consistent, easily parseable format for downstream applications.

  • Identifying claims relies on linguistic cues separate from truthfulness, and the study assumes models should do the same; observed safety behaviors and genre effects suggest that even with similar training data, LLMs still conflate real vs. fake content.

Paper 3: Background

Paper 3: Set up

Causal Sequence Classification

Causal Span Detection

Paper 3: Training specifications

Corpus

  • AJ, BBC and CNN
  • May 2023 to February 2024
Region Source Count (Proportion)
ME AJ 1251 (0.48)
BBC 792 (0.30)
CNN 567 (0.22)
EE AJ 1018 (0.43)
BBC 784 (0.33)
CNN 581 (0.24)

Training splits

Set Count (Proportion)
Train 862 (0.70)
Test 191 (0.15)
Validation 185 (0.15)
Total 1238 (1.00)

Datasets

🤗 pgarco/CausalPalUkr_Train_Test

🤗 pgarco/CausalPalUkr_Val.

Models

🤗 pgarco/CausalPalUkr_token

🤗 pgarco/CausalPalUkr_seq

Model metrics

Precision Recall Accuracy F1-score
CAUSAL 0.891 0.855 0.872
NOT CAUSAL 0.727 0.787 0.756
weighted average 0.837 0.832 0.832 0.834

Paper 3: Validation

Paper 3: Results

Paper 3: Discussion

  • Model Selection: Model agnostic and potential to include LLMs
  • Annotation rules can be further detailed for specific cases
  • Additional information due to a known semantic relationship

Paper 4: Background

Paper 4: Operationalization

Causal with responsibility Causal without responsibility
Officials report indicent in Kharkiv. Officials report incident in Kharkiv kills 20.


Agentless with overall ambiguity Agentless with precise outcome
Officials report incdent in Kharkic. Officials report incident in Kharkiv kills 20.


Non-attribution by supression Non-attribution by predicate change
20 killed in Kharkiv strike. Officials in Kharkic report 20 dead.

Paper 4: Corpus and data

  • AJ, BBC and CNN
  • October 2023 to June 2025
Region Source Count (Proportion)
ME AJ 1994 (0.29)
BBC 2781 (0.41)
CNN 2017 (0.30)
EE AJ 3603 (0.39)
BBC 2625 (0.29)
CNN 2962 (0.32)

Paper 4: Agent Annotation

Paper 4: Results







Paper 4: Discussion

  • Low used tokens render very big CI

  • News events are not matched nor weighted

  • NER might be more robust than “actor = subject” analysis

  • Mechanism might be a function of maximizing content production and minimizing ideological inconsistency

Bibliography

Bisgaard, Martin. 2015. “Bias Will Find a Way: Economic Perceptions, Attributions of Blame, and Partisan-Motivated Reasoning During Crisis.” The Journal of Politics 77 (3): 849–60. https://doi.org/10.1086/681591.
Caselli, Tommaso, and Piek Vossen. 2017. “Proceedings of the Events and Stories in the News Workshop.” In, 77–86. Vancouver, Canada: Association for Computational Linguistics. https://doi.org/10.18653/v1/W17-2711.
Corral, Paulina Garcia, Hanna Bechara, Ran Zhang, and Slava Jankin. 2024. “PolitiCause: An Annotation Scheme and Corpus for Causality in Political Texts.” Freie Universität Berlin. https://doi.org/10.17169/REFUBIUM-45757.
Devlin, Jacob, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. “BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding.” https://doi.org/10.48550/arXiv.1810.04805.
Drury, Brett, Hugo Gonçalo Oliveira, and Alneu de Andrade Lopes. 2022. “A Survey of the Extraction and Applications of Causal Relations.” Natural Langauge Engineering. https://doi.org/doi:10.1017/S135132492100036X.
Du, Li, Xiao Ding, Kai Xiong, Ting Liu, and Bing Qin. 2022. “ACL 2022.” In, 432446. Dublin, Ireland: Association for Computational Linguistics. https://doi.org/10.18653/v1/2022.acl-long.33.
Dunietz, Jesse. n.d. “Annotating and Automatically Tagging Constructions of Causal Language,” 207.
Entman, Robert M. 1993. “Framing: Toward Clarification of a Fractured Paradigm.” Journal of Communication 43 (4): 51–58. https://doi.org/10.1111/j.1460-2466.1993.tb01304.x.
Gao, Jinglong, Xiao Ding, Bing Qin, and Ting Liu. 2023. “Is ChatGPT a Good Causal Reasoner? A Comprehensive Evaluation.” In Findings of the Association for Computational Linguistics: EMNLP 2023, edited by Houda Bouamor, Juan Pino, and Kalika Bali, 11111–26. Singapore: Association for Computational Linguistics. https://doi.org/10.18653/v1/2023.findings-emnlp.743.
Gentzkow, Matthew, and Jesse M. Shapiro. 2006. “Media Bias and Reputation.” Journal of Political Economy 114 (2): 280–316. https://doi.org/10.1086/499414.
Graham, Matthew H., and Shikhar Singh. 2023. “An Outbreak of Selective Attribution: Partisanship and Blame in the COVID-19 Pandemic.” American Political Science Review 118 (1): 423–41. https://doi.org/10.1017/s0003055423000047.
Grose, Christian R., Neil Malhotra, and Robert Parks Van Houweling. 2015. “Explaining Explanations: How Legislators Explain Their Policy Positions and How Citizens React.” American Journal of Political Science 59: 724–43. https://doi.org/https://doi.org/10.1111/ajps.12164.
Gu, Jinghang, Longhua Qian, and Guodong Zhou. 2016. “Chemical-Induced Disease Relation Extraction with Various Linguistic Features.” Database, 42. https://doi.org/10.1093/database/baw042.
Hameleers, Michael, Linda Bos, and Claes H. de Vreese. 2017. They Did It: The Effects of Emotionalized Blame Attribution in Populist Communication.” Communication Research 44 (6): 870–900. https://doi.org/10.1177/0093650216644026.
Hansson, Sten. 2015. “Discursive Strategies of Blame Avoidance in Government: A Framework for Analysis.” Discourse & Society 26 (3): 297–322. https://doi.org/10.1177/0957926514564736.
Hidey, Christopher, and Kathy McKeown. 2016. “ACL 2016.” In, 14241433. Berlin, Germany: Association for Computational Linguistics. https://doi.org/10.18653/v1/P16-1135.
Ho, Matthew, Aditya Sharma, Justin Chang, Michael Saxon, Sharon Levy, Yujie Lu, and William Yang Wang. 2022. “WikiWhy: Answering and Explaining Cause-and-Effect Questions.” https://arxiv.org/abs/2210.12152.
Hobbhahn, Marius, Tom Lieberum, and David Seiler. 2022. “Investigating Causal Understanding in LLMs.” In NeurIPS 2022 Workshop on Causality for Real-World Impact. https://openreview.net/forum?id=st6jtGdW8Ke.
Hobolt, Sara B., and James Tilley. 2014. “Blaming Europe?” February. https://doi.org/10.1093/acprof:oso/9780199665686.001.0001.
Iyengar, Shanto. 1987. “Television News and Citizens’ Explanations of National Affairs.” The American Political Science Review 81 (3): 815–31. https://doi.org/10.2307/1962678.
———. 1989. “How Citizens Think about National Issues: A Matter of Responsibility.” American Journal of Political Science 33 (4): 878. https://doi.org/10.2307/2111113.
Jankin, Slava, Alexander Baturo, and Niheer Dasandi. 2017. “United Nations General Debate Corpus.” https://doi.org/10.7910/DVN/0TJX8Y.
Jin, Zhijing, Jiarui Liu, Zhiheng Lyu, Spencer Poff, Mrinmaya Sachan, Rada Mihalcea, Mona Diab, and Bernhard Schölkopf. 2024. “Can Large Language Models Infer Causation from Correlation?” https://arxiv.org/abs/2306.05836.
Kim, Sei-Hill. 2015. “Who Is Responsible for a Social Problem? News Framing and Attribution of Responsibility.” Journalism & Mass Communication Quarterly 92 (3): 554–58. https://doi.org/10.1177/1077699015591956.
Kıcıman, Emre, Robert Ness, Amit Sharma, and Chenhao Tan. 2023. “Causal Reasoning and Large Language Models: Opening a New Frontier for Causality.” arXiv. https://doi.org/10.48550/arXiv.2305.00050.
Malhotra, Neil, and Alexander G. Kuo. 2008. “Attributing Blame: The Public’s Response to Hurricane Katrina.” The Journal of Politics 70 (1): 120–35. https://doi.org/10.1017/s0022381607080097.
Mariko, Dominique, Hanna Aki Akl, Estelle Labidurie, Stephane Durfort, Hugues de Mazancourt, and Mahmoud El-Haj. n.d. “The Financial Document Causality Detection Shared Task (FinCausal 2021).”
McCabe, Katherine T. 2016. “Attitude Responsiveness and Partisan Bias: Direct Experience with the Affordable Care Act.” Political Behavior 38: 861–82. https://doi.org/10.1007/s11109-016-9337-9.
Mirza, Paramita, Rachele Sprugnoli, Sara Tonelli, and Manuela Speranza. 2014. “Annotating Causality in the TempEval-3 Corpus.” In, 1019. Gothenburg, Sweden: Association for Computational Linguistics. https://doi.org/10.3115/v1/W14-0702.
Rozenas, Arturas, and Denis Stukal. 2019. “How Autocrats Manipulate Economic News: Evidence from Russia’s State-Controlled Television.” The Journal of Politics 81 (3): 982–96. https://doi.org/10.1086/703208.
Rudolph, Thomas J. 2006. “Triangulating Political Responsibility: The Motivated Formation of Responsibility Judgments.” Political Psychology 27 (1): 99–122. https://doi.org/https://doi.org/10.1111/j.1467-9221.2006.00451.x.
Sahar, Gail. 2014. “On the Importance of Attribution Theory in Political Psychology.” Social and Personality Psychology Compass 8 (5): 229–49. https://doi.org/10.1111/spc3.12102.
Shiller, Robert J. 2019. Narrative Economics: How Stories Go Viral and Drive Major Economic Events. Princeton University Press. https://ebookcentral.proquest.com/lib/fuberlin-ebooks/detail.action?pq-origsite=primo&docID=5847663.
Sotirovic, Mira. 2003. “How Individuals Explain Social Problems: The Influences of Media Use.” Journal of Communication.
Stone, Deborah A. 1989. “Causal Stories and the Formation of Policy Agendas.” Political Science Quarterly 104 (2): 281–300. https://doi.org/10.2307/2151585.
Takayanagi, Takehiro, Masahiro Suzuki, Ryotaro Kobayashi, Hiroki Sakaji, and Kiyoshi Izumi. 2024. “Is ChatGPT the Future of Causal Text Mining? A Comprehensive Evaluation and Analysis.” https://arxiv.org/abs/2402.14484.
Tan, Fiona Anting, Ali Hürriyetoğlu, Tommaso Caselli, Nelleke Oostdijk, Tadashi Nomoto, Hansi Hettiarachchi, Iqra Ameer, Onur Uca, Farhana Ferdousi Liza, and Tiancheng Hu. 2022. “The Causal News Corpus: Annotating Causal Relations in Event Sentences from News.” In Proceedings of the Thirteenth Language Resources and Evaluation Conference, edited by Nicoletta Calzolari, Frédéric Béchet, Philippe Blache, Khalid Choukri, Christopher Cieri, Thierry Declerck, Sara Goggi, et al., 2298–2310. Marseille, France: European Language Resources Association. https://aclanthology.org/2022.lrec-1.246/.
Tan, Fiona Anting, Xinyu Zuo, and See-Kiong Ng. 2023. “UniCausal: Unified Benchmark and Repository for Causal Text Mining.”
Tappin, Ben M., Gordon Pennycook, and David G. Rand. 2021. “Rethinking the Link Between Cognitive Sophistication and Politically Motivated Reasoning.” Journal of Experimental Psychology: General 150 (6): 1095–1114. https://doi.org/10.1037/xge0000974.
Tilley, James, and Sara B. Hobolt. 2011. “Is the Government to Blame? An Experimental Test of How Partisanship Shapes Perceptions of Performance and Responsibility.” The Journal of Politics 73 (2): 316–30. https://doi.org/10.1017/s0022381611000168.
Yu, Bei, Yingya Li, and Jun Wang. n.d. “Detecting Causal Language Use in Science Findings.”