APS Logo

QuanTA: Efficient High-Rank Fine-Tuning of LLMs with Quantum-Informed Tensor Adaptation

ORAL

Abstract

Large Language Models (LLMs) have become critical tools across many domains, but fine-tuning them for specific tasks remains a challenge. We propose Quantum-informed Tensor Adaptation (QuanTA), a novel method that utilizes quantum computation-inspired techniques to achieve efficient high-rank fine-tuning. Unlike Low-Rank Adaptation (LoRA), which may struggle with complex downstream tasks due to its low-rank nature, QuanTA leverages high-rank adaptations, supported by theoretical results such as the universality and rank representation theorems, to overcome these limitations. QuanTA offers improved performance in commonsense reasoning, arithmetic reasoning, and scalability while using fewer trainable parameters than other methods, without introducing inference overhead. Furthermore, QuanTA can be integrated with existing fine-tuning algorithms, providing a scalable and efficient approach to enhancing LLMs and advancing state-of-the-art in natural language processing.

Publication: https://nips.cc/virtual/2024/poster/96019<br>https://arxiv.org/abs/2406.00132

Presenters

  • Zhuo Chen

    Massachusetts Institute of Technology

Authors

  • Zhuo Chen

    Massachusetts Institute of Technology

  • Rumen Dangovski

    Massachusetts Institute of Technology

  • Charlotte Loh

    Massachusetts Institute of Technology

  • Owen Dugan

    Massachusetts Institute of Technology

  • Di Luo

    Massachusetts Institute of Technology

  • Marin Soljačić

    Massachusetts Institute of Technology