
Why This Matters
Political scientists rely on reading political texts to infer party positions, policy priorities, and likely coalition behavior, but human coding is costly and slow while automated text-as-data methods rest on strong assumptions. Kenneth Benoit, Scott De Marchi, Conor Laver, Michael Laver, and Jinshuai Ma test whether large language models (LLMs) can provide a scalable, interpretable middle ground by performing Natural Language Understanding (NLU) on political documents.
What the Authors Do
The authors develop a systematic, replicable workflow that uses ensembles of LLM outputs to interpret political texts as meaningful statements about actors and issues rather than treating text purely as token counts. They apply this method to estimate party positions on six key issue dimensions and to classify content of coalition policy declarations.
How the Test Was Set Up
Key Findings
Implications for Political Science
Benoit et al. argue that modern LLMs can reduce the trade-off between the depth of human qualitative coding and the scalability of statistical text methods. Their results suggest LLMs can reliably approximate expert judgment and may improve measurement for research on party positions and coalition formation. The paper concludes with a discussion of methodological opportunities and cautionary notes about limitations and future validation needs.

| Using Large Language Models to Analyze Political Texts through Natural Language Understanding was authored by Kenneth Benoit, Scott De Marchi, Conor Laver, Michael Laver and Jinshuai Ma. It was published by Wiley in AJPS in 2025. |