Combining Induction and Transduction for Abstract Reasoning
When learning an input-output mapping from very few examples, is it better to first infer a latent function that explains the examples, or is it better to directly predict new test outputs, e.g. using a neural network? We study this question on ARC by training neural models for induction (inferring...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | When learning an input-output mapping from very few examples, is it better to
first infer a latent function that explains the examples, or is it better to
directly predict new test outputs, e.g. using a neural network? We study this
question on ARC by training neural models for induction (inferring latent
functions) and transduction (directly predicting the test output for a given
test input). We train on synthetically generated variations of Python programs
that solve ARC training tasks. We find inductive and transductive models solve
different kinds of test problems, despite having the same training problems and
sharing the same neural architecture: Inductive program synthesis excels at
precise computations, and at composing multiple concepts, while transduction
succeeds on fuzzier perceptual concepts. Ensembling them approaches human-level
performance on ARC. |
---|---|
DOI: | 10.48550/arxiv.2411.02272 |