Content Quality vs. Attention Allocation: An LLM-Based Case Study in Peer-to-peer Mental Health Networks
With the rise of social media and peer-to-peer networks, users increasingly rely on crowdsourced responses for information and assistance. However, the mechanisms used to rank and promote responses often prioritize and end up biasing in favor of timeliness over quality, which may result in suboptima...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | With the rise of social media and peer-to-peer networks, users increasingly
rely on crowdsourced responses for information and assistance. However, the
mechanisms used to rank and promote responses often prioritize and end up
biasing in favor of timeliness over quality, which may result in suboptimal
support for help-seekers. We analyze millions of responses to mental
health-related posts, utilizing large language models (LLMs) to assess the
multi-dimensional quality of content, including relevance, empathy, and
cultural alignment, among other aspects. Our findings reveal a mismatch between
content quality and attention allocation: earlier responses - despite being
relatively lower in quality - receive disproportionately high fractions of
upvotes and visibility due to platform ranking algorithms. We demonstrate that
the quality of the top-ranked responses could be improved by up to 39 percent,
and even the simplest re-ranking strategy could significantly improve the
quality of top responses, highlighting the need for more nuanced ranking
mechanisms that prioritize both timeliness and content quality, especially
emotional engagement in online mental health communities. |
---|---|
DOI: | 10.48550/arxiv.2411.05328 |