Text2Traj2Text: Learning-by-Synthesis Framework for Contextual Captioning of Human Movement Trajectories
This paper presents Text2Traj2Text, a novel learning-by-synthesis framework for captioning possible contexts behind shopper's trajectory data in retail stores. Our work will impact various retail applications that need better customer understanding, such as targeted advertising and inventory ma...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | This paper presents Text2Traj2Text, a novel learning-by-synthesis framework
for captioning possible contexts behind shopper's trajectory data in retail
stores. Our work will impact various retail applications that need better
customer understanding, such as targeted advertising and inventory management.
The key idea is leveraging large language models to synthesize a diverse and
realistic collection of contextual captions as well as the corresponding
movement trajectories on a store map. Despite learned from fully synthesized
data, the captioning model can generalize well to trajectories/captions created
by real human subjects. Our systematic evaluation confirmed the effectiveness
of the proposed framework over competitive approaches in terms of ROUGE and
BERT Score metrics. |
---|---|
DOI: | 10.48550/arxiv.2409.12670 |