Missed Connections: Lateral Thinking Puzzles for Large Language Models
The Connections puzzle published each day by the New York Times tasks players with dividing a bank of sixteen words into four groups of four words that each relate to a common theme. Solving the puzzle requires both common linguistic knowledge (i.e. definitions and typical usage) as well as, in many...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The Connections puzzle published each day by the New York Times tasks players
with dividing a bank of sixteen words into four groups of four words that each
relate to a common theme. Solving the puzzle requires both common linguistic
knowledge (i.e. definitions and typical usage) as well as, in many cases,
lateral or abstract thinking. This is because the four categories ascend in
complexity, with the most challenging category often requiring thinking about
words in uncommon ways or as parts of larger phrases. We investigate the
capacity for automated AI systems to play Connections and explore the game's
potential as an automated benchmark for abstract reasoning and a way to measure
the semantic information encoded by data-driven linguistic systems. In
particular, we study both a sentence-embedding baseline and modern large
language models (LLMs). We report their accuracy on the task, measure the
impacts of chain-of-thought prompting, and discuss their failure modes.
Overall, we find that the Connections task is challenging yet feasible, and a
strong test-bed for future work. |
---|---|
DOI: | 10.48550/arxiv.2404.11730 |