A Novel Adaptive Fine-Tuning Algorithm for Multimodal Models: Self-Optimizing Classification and Selection of High-Quality Datasets in Remote Sensing
We propose an adaptive fine-tuning algorithm for multimodal large models. The core steps of this algorithm involve two stages of truncation. First, the vast amount of data is projected into a semantic vector space, and the MiniBatchKMeans algorithm is used for automated clustering. This classificati...
Gespeichert in:
Hauptverfasser: | , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | We propose an adaptive fine-tuning algorithm for multimodal large models. The
core steps of this algorithm involve two stages of truncation. First, the vast
amount of data is projected into a semantic vector space, and the
MiniBatchKMeans algorithm is used for automated clustering. This classification
ensures that the data within each cluster exhibit high semantic similarity.
Next, we process the data in each cluster, calculating the translational
difference between the original and perturbed data in the multimodal large
model's vector space. This difference serves as a generalization metric for the
data. Based on this metric, we select the data with high generalization
potential for training. We applied this algorithm to train the
InternLM-XComposer2-VL-7B model on two 3090 GPUs using one-third of the GeoChat
multimodal remote sensing dataset. The results demonstrate that our algorithm
outperforms the state-of-the-art baselines. various baselines. The model
trained on our optimally chosen one-third dataset, based on experimental
validation, exhibited only 1% reduction in performance across various remote
sensing metrics compared to the model trained on the full dataset. This
approach significantly preserved general-purpose capabilities while reducing
training time by 68.2%. Furthermore, the model achieved scores of 89.86 and
77.19 on the UCMerced and AID evaluation datasets, respectively, surpassing the
GeoChat dataset by 5.43 and 5.16 points. It only showed a 0.91-point average
decrease on the LRBEN evaluation dataset. |
---|---|
DOI: | 10.48550/arxiv.2409.13345 |