Back to Search Start Over

Efficient Multi-modal Large Language Models via Visual Token Grouping

Authors :
Huang, Minbin
Huang, Runhui
Shi, Han
Chen, Yimeng
Zheng, Chuanyang
Sun, Xiangguo
Jiang, Xin
Li, Zhenguo
Cheng, Hong
Publication Year :
2024

Abstract

The development of Multi-modal Large Language Models (MLLMs) enhances Large Language Models (LLMs) with the ability to perceive data formats beyond text, significantly advancing a range of downstream applications, such as visual question answering and image captioning. However, the substantial computational costs associated with processing high-resolution images and videos pose a barrier to their broader adoption. To address this challenge, compressing vision tokens in MLLMs has emerged as a promising approach to reduce inference costs. While existing methods conduct token reduction in the feature alignment phase. In this paper, we introduce VisToG, a novel grouping mechanism that leverages the capabilities of pre-trained vision encoders to group similar image segments without the need for segmentation masks. Specifically, we concatenate semantic tokens to represent image semantic segments after the linear projection layer before feeding into the vision encoder. Besides, with the isolated attention we adopt, VisToG can identify and eliminate redundant visual tokens utilizing the prior knowledge in the pre-trained vision encoder, which effectively reduces computational demands. Extensive experiments demonstrate the effectiveness of VisToG, maintaining 98.1% of the original performance while achieving a reduction of over 27\% inference time.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2411.17773
Document Type :
Working Paper