Towards Human-Level Text Coding with LLMs: The Case of Fatherhood Roles in Public Policy Documents
Recent advances in large language models (LLMs) like GPT-3.5 and GPT-4 promise automation with better results and less programming, opening up new opportunities for text analysis in political science. In this study, we evaluate LLMs on three original coding tasks involving typical complexities encou...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Recent advances in large language models (LLMs) like GPT-3.5 and GPT-4
promise automation with better results and less programming, opening up new
opportunities for text analysis in political science. In this study, we
evaluate LLMs on three original coding tasks involving typical complexities
encountered in political science settings: a non-English language, legal and
political jargon, and complex labels based on abstract constructs. Along the
paper, we propose a practical workflow to optimize the choice of the model and
the prompt. We find that the best prompting strategy consists of providing the
LLMs with a detailed codebook, as the one provided to human coders. In this
setting, an LLM can be as good as or possibly better than a human annotator
while being much faster, considerably cheaper, and much easier to scale to
large amounts of text. We also provide a comparison of GPT and popular
open-source LLMs, discussing the trade-offs in the model's choice. Our software
allows LLMs to be easily used as annotators and is publicly available:
https://github.com/lorelupo/pappa. |
---|---|
DOI: | 10.48550/arxiv.2311.11844 |