MAO: Efficient Model-Agnostic Optimization of Prompt Tuning for Vision-Language Models

March 23, 2025 ยท Declared Dead ยท ๐Ÿ› IEEE International Conference on Multimedia and Expo

๐Ÿ’€ CAUSE OF DEATH: 404 Not Found
Code link is broken/dead
Authors Haoyang Li, Siyu Zhou, Liang Wang, Guodong Long arXiv ID 2503.18160 Category cs.CV: Computer Vision Cross-listed cs.MM Citations 0 Venue IEEE International Conference on Multimedia and Expo Repository https://github.com/JREion/M.A.O Last Checked 2 months ago
Abstract
Though CLIP-based prompt tuning significantly enhances pre-trained Vision-Language Models, existing research focuses on reconstructing the model architecture, e.g., additional loss calculation and meta-networks. These approaches generally lead to increased complexity and extended training cost. To maintain the efficiency of the tuning process, we propose plug-and-play Model-Agnostic Optimization (MAO) for prompt tuning. Without altering any components of the prompt tuning backbone, we introduce a Data-Driven Enhancement framework to optimize the distribution of the initial data, and incorporate an Alterable Regularization module to boost the task-specific feature processing pipeline, thereby improving overall performance while maintaining low computational cost. Extensive experiments on MAO demonstrate its outstanding performance and efficiency. The code of MAO is available at: https://github.com/JREion/M.A.O .
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision

Died the same way โ€” ๐Ÿ’€ 404 Not Found