Are Foundation Models the Next-Generation Social Media Content Moderators?
Recent progress in artificial intelligence (AI) tools and systems has been significant, especially in their reasoning and efficiency. Notable examples include generative AI-based large language models (LLMs) like Generative Pre-trained Transformer 3.5 (GPT-3.5), GPT-4, and Gemini, among others. In o...
Gespeichert in:
Veröffentlicht in: | IEEE intelligent systems 2024-11, Vol.39 (6), p.70-80 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Recent progress in artificial intelligence (AI) tools and systems has been significant, especially in their reasoning and efficiency. Notable examples include generative AI-based large language models (LLMs) like Generative Pre-trained Transformer 3.5 (GPT-3.5), GPT-4, and Gemini, among others. In our work, we evaluated the effectiveness of fine-tuned deep learning models compared to general-purpose LLMs in moderating image-based content. We used deep learning models such as convolutional neural networks, ResNet50, and VGG-16, trained them for violence detection on an image dataset, and tested them on a separate dataset. The same test dataset was also evaluated using Large Language and Vision Assistant (LLaVa) and GPT-4, two LLMs that can process images. The results demonstrate that VGG-16 model had the highest accuracy at 0.94, while LLaVa had the lowest at 0.66. GPT-4 showed superiority over LLaVa with an accuracy value of 0.9242. LLaVa recorded the highest precision of all models. |
---|---|
ISSN: | 1541-1672 |
DOI: | 10.1109/MIS.2024.3477109 |