loading page

Efficient Biomedical Text Summarization with Quantized LLAMA 2: Enhancing Memory Usage and Inference on Low Powered Devices
  • +2
  • Sanjeev Kumar,
  • Vikas Ranjan,
  • Arjab Chakrabarti,
  • Tridib Kumar Das,
  • Anushka Singh
Sanjeev Kumar
University of Illinois Urbana-Champaign Institute of Communications Research

Corresponding Author:[email protected]

Author Profile
Vikas Ranjan
Wayfair LLC
Author Profile
Arjab Chakrabarti
Kalinga Institute of Industrial Technology Deemed to be University
Author Profile
Tridib Kumar Das
Kalinga Institute of Industrial Technology Deemed to be University
Author Profile
Anushka Singh
Kalinga Institute of Industrial Technology Deemed to be University
Author Profile

Abstract

The deployment of large language models (LLMs) on edge devices and non-server environments presents significant challenges, primarily due to constraints in memory usage, computational power, and inference time. This paper investigates the feasibility of running LLMs across such devices by focusing on optimizing memory usage, employing quantization techniques, and reducing inference time. Specifically, we utilize LLaMA 2 for biomedical text summarization and implement Low-Rank Adaptation (LoRA) quantization to compress the model size to compress the model size and fine-tune it using limited resources. Our study systematically evaluates memory consumption during both training and inference phases, demonstrating substantial reductions through efficient LoRA quantization. Our results indicate that with careful optimization, it is feasible to deploy sophisticated LLMs like LLaMA 2 on low powered devices, thereby broadening the scope of their application in resource-constrained environments.
Submitted to Expert Systems
Submission Checks Completed
Assigned to Editor
Reviewer(s) Assigned