TY - JOUR
T1 - Towards algorithmic framing analysis
T2 - expanding the scope by using LLMs
AU - Kuang, Xianwen
AU - Liu, Jun
AU - Zhang, Haiyang
AU - Schweighofer, Simon
N1 - Publisher Copyright:
© The Author(s) 2025.
PY - 2025/12
Y1 - 2025/12
N2 - Framing analysis, an extensively used, multi-disciplinary social science research method, requires substantial manpower and time to code and uncover human-level understanding of story contexts. However, recent advances in deep learning have led to a qualitative jump in algorithm-assisted methods, with large language models (LLMs) like BERT and GPT going beyond surface characteristics to infer the semantic properties of a text. In this study, we explore the application of the BERT for natural language inference (NLI), which leverages bidirectional context and rich embeddings to assist scholars in identifying contextual information in media texts for quantitative framing analysis. More specifically, we investigate the capability of LLMs to identify generic media frames by comparing the results from a zero-shot analysis using BERT-NLI to those from human analysis. We find that the reliability of detecting generic frames varies significantly across different datasets, indicating that even a large LLM like BERT-NLI, trained on millions of texts from diverse sources, cannot be uniformly trusted across different contexts. Nonetheless, LLMs might be employed productively in specific contexts after careful consideration of their agreement with human-generated ratings.
AB - Framing analysis, an extensively used, multi-disciplinary social science research method, requires substantial manpower and time to code and uncover human-level understanding of story contexts. However, recent advances in deep learning have led to a qualitative jump in algorithm-assisted methods, with large language models (LLMs) like BERT and GPT going beyond surface characteristics to infer the semantic properties of a text. In this study, we explore the application of the BERT for natural language inference (NLI), which leverages bidirectional context and rich embeddings to assist scholars in identifying contextual information in media texts for quantitative framing analysis. More specifically, we investigate the capability of LLMs to identify generic media frames by comparing the results from a zero-shot analysis using BERT-NLI to those from human analysis. We find that the reliability of detecting generic frames varies significantly across different datasets, indicating that even a large LLM like BERT-NLI, trained on millions of texts from diverse sources, cannot be uniformly trusted across different contexts. Nonetheless, LLMs might be employed productively in specific contexts after careful consideration of their agreement with human-generated ratings.
KW - BERT-NLI
KW - Big Data
KW - Context
KW - Framing analysis
KW - Large language models (LLMs)
UR - http://www.scopus.com/inward/record.url?scp=105000055782&partnerID=8YFLogxK
U2 - 10.1186/s40537-025-01092-y
DO - 10.1186/s40537-025-01092-y
M3 - Article
AN - SCOPUS:105000055782
SN - 2196-1115
VL - 12
JO - Journal of Big Data
JF - Journal of Big Data
IS - 1
M1 - 66
ER -