AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
Article Link
Collect
Submit Manuscript
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Regular Paper

KnowBench: Evaluating the Knowledge Alignment on Large Visual Language Models

National Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210023, China
School of Artificial Intelligence, Nanjing University, Nanjing 210023, China

Equal Contributed (Zheng Ma proposed the overall research concept, provided the original data, and drafted the manuscript; Hao-Tian Yang took charge of data generation and model testing, and created the figures in the paper.)

Show Author Information

Abstract

Large visual language models (LVLMs) have revolutionized the multimodal domain, demonstrating exceptional performance in tasks requiring fusing visual and textual information. However, the current evaluation benchmarks fail to adequately assess the knowledge alignment between images and text, focusing primarily on answer accuracy rather than the reasoning processes behind them. To address this gap and enhance the understanding of LVLMs’ capabilities, we introduce KnowBench, a novel benchmark designed to assess the alignment of knowledge between images and text for LVLMs. KnowBench comprises 1 081 image-question pairs, each with four options and four pieces of corresponding knowledge across 11 major categories. We evaluate mainstream LVLMs on KnowBench, including proprietary models like Gemini, Claude, and GPT, and open-source models like LLaVA, Qwen-VL, and InternVL. Our experiments reveal a notable discrepancy in the models’ abilities to select correct answers and corresponding knowledge whether the models are open-source or proprietary. This indicates that there is still a significant gap in the current LVLMs’ knowledge alignment between images and text. Furthermore, our further analysis shows that model performance on KnowBench improves with increased parameters and version iterations. This indicates that scaling laws have a significant impact on multimodal knowledge alignment, and the iteration of the model by researchers also has a positive effect. We anticipate that KnowBench will foster the development of LVLMs and motivate researchers to develop more reliable models. We have made our dataset publicly available at https://doi.org/10.57760/sciencedb.29672.

Electronic Supplementary Material

Download File(s)
JCST-2504-15512-Highlights.pdf (856.7 KB)

References

【1】
【1】
 
 
Journal of Computer Science and Technology
Pages 1209-1219

{{item.num}}

Comments on this article

Go to comment

< Back to all reports

Review Status: {{reviewData.commendedNum}} Commended , {{reviewData.revisionRequiredNum}} Revision Required , {{reviewData.notCommendedNum}} Not Commended Under Peer Review

Review Comment

Close
Close
Cite this article:
Ma Z, Yang H-T, Zhang J-B, et al. KnowBench: Evaluating the Knowledge Alignment on Large Visual Language Models. Journal of Computer Science and Technology, 2025, 40(5): 1209-1219. https://doi.org/10.1007/s11390-025-5512-y

496

Views

0

Crossref

0

Web of Science

0

Scopus

0

CSCD

Received: 30 April 2025
Accepted: 08 September 2025
Published: 10 September 2025
© Institute of Computing Technology, Chinese Academy of Sciences 2025