Skip to content

IJCAI 2025, FBQuant: FeedBack Quantization for Large Language Models

Notifications You must be signed in to change notification settings

kriskrisliu/FeedBackQuant

Repository files navigation

FBQuant: FeedBack Quantization for Large Language Models

Accepted to IJCAI 2025. Paper link.


Authors

Yijiang Liu, Hengyu Fang, Liulu He, Rongyu Zhang, Yichuan Bai, Yuan Du, Li Du

Authors are from Nanjing University


Overview

Deploying Large Language Models (LLMs) on edge devices is increasingly important, as it eliminates reliance on network connections, reduces expensive API calls, and enhances user privacy. However, on-device deployment is challenging due to the limited computational resources of edge devices. In particular, the key bottleneck stems from memory bandwidth constraints related to weight loading. Weight-only quantization effectively reduces memory access, yet often induces significant accuracy degradation. Recent efforts to incorporate sub-branches have shown promise for mitigating quantization errors, but these methods either lack robust optimization strategies or rely on suboptimal objectives. To address these gaps, we propose FeedBack Quantization (FBQuant), a novel approach inspired by negative feedback mechanisms in automatic control. FBQuant inherently ensures that the reconstructed weights remain bounded by the quantization process, thereby reducing the risk of overfitting. To further offset the additional latency introduced by sub-branches, we develop an efficient CUDA kernel that decreases 60% of extra inference time. Comprehensive experiments demonstrate the efficiency and effectiveness of FBQuant across various LLMs. Notably, for 3-bit Llama2-7B, FBQuant improves zero-shot accuracy by 1.2%.


Checklist

  • Clean the codebase: Ensure the code is organized, readable, and optimized.
  • Write detailed instructions: Provide a clear guide for running the program.
  • Open-source the CUDA Kernel: Release the efficient CUDA kernel developed for reducing inference latency.

Acknowledgment

We extend our gratitude to the following works that inspired and supported this research:

  • GPTQ
  • AWQ
  • OmniQuant

Citation

@article{liu2025fbquant,
  title={FBQuant: FeedBack Quantization for Large Language Models},
  author={Liu, Yijiang and Fang, Hengyu and He, Liulu and Zhang, Rongyu and Bai, Yichuan and Du, Yuan and Du, Li},
  journal={arXiv preprint arXiv:2501.16385},
  year={2025}
}

About

IJCAI 2025, FBQuant: FeedBack Quantization for Large Language Models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published