SuperSAM: Crafting a SAM Supernetwork via Structured Pruning and Unstructured Parameter Prioritization
Neural Architecture Search (NAS) is a powerful approach of automating the design of efficient neural architectures. In contrast to traditional NAS methods, recently proposed one-shot NAS methods prove to be more efficient in performing NAS. One-shot NAS works by generating a singular weight-sharing...
Gespeichert in:
Hauptverfasser: | , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Neural Architecture Search (NAS) is a powerful approach of automating the
design of efficient neural architectures. In contrast to traditional NAS
methods, recently proposed one-shot NAS methods prove to be more efficient in
performing NAS. One-shot NAS works by generating a singular weight-sharing
supernetwork that acts as a search space (container) of subnetworks. Despite
its achievements, designing the one-shot search space remains a major
challenge. In this work we propose a search space design strategy for Vision
Transformer (ViT)-based architectures. In particular, we convert the Segment
Anything Model (SAM) into a weight-sharing supernetwork called SuperSAM. Our
approach involves automating the search space design via layer-wise structured
pruning and parameter prioritization. While the structured pruning applies
probabilistic removal of certain transformer layers, parameter prioritization
performs weight reordering and slicing of MLP-blocks in the remaining layers.
We train supernetworks on several datasets using the sandwich rule. For
deployment, we enhance subnetwork discovery by utilizing a program autotuner to
identify efficient subnetworks within the search space. The resulting
subnetworks are 30-70% smaller in size compared to the original pre-trained SAM
ViT-B, yet outperform the pretrained model. Our work introduces a new and
effective method for ViT NAS search-space design. |
---|---|
DOI: | 10.48550/arxiv.2501.08504 |