Rethinking Compression: Reduced Order Modelling of Latent Features in Large Language Models

December 12, 2023 ยท Declared Dead ยท ๐Ÿ› arXiv.org

๐Ÿ“œ CAUSE OF DEATH: Death by README
Repo has only a README

Repo contents: .gitignore, LICENSE, README.md

Authors Arnav Chavan, Nahush Lele, Deepak Gupta arXiv ID 2312.07046 Category cs.LG: Machine Learning Cross-listed cs.CL Citations 1 Venue arXiv.org Repository https://github.com/transmuteAI/trailmet/tree/main/trailmet/algorithms/llm-rom โญ 19 Last Checked 1 month ago
Abstract
Due to the substantial scale of Large Language Models (LLMs), the direct application of conventional compression methodologies proves impractical. The computational demands associated with even minimal gradient updates present challenges, particularly on consumer-grade hardware. This paper introduces an innovative approach for the parametric and practical compression of LLMs based on reduced order modelling, which entails low-rank decomposition within the feature space and re-parameterization in the weight space. Notably, this compression technique operates in a layer-wise manner, obviating the need for a GPU device and enabling the compression of billion-scale models within stringent constraints of both memory and time. Our method represents a significant advancement in model compression by leveraging matrix decomposition, demonstrating superior efficacy compared to the prevailing state-of-the-art structured pruning method.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning

Died the same way โ€” ๐Ÿ“œ Death by README