Benchmarking and Improving Detail Image Caption (2024)

Hongyuan Dong1*,Jiawen Li1*,Bohong Wu1,Jiacong Wang1,2,Yuan Zhang1,3,Haoyuan Guo1
1ByteDance Inc.  2School of Artificial Intelligence, University of Chinese Academy of Sciences
3School of Computer Science, Peking University
{donghongyuan.dousia, lijiawen.0818, bohongwu}@bytedance.com
wangjiacong20@mails.ucas.ac.cn, {zhangyuan.gump, guohaoyuan}@bytedance.com

Abstract

Image captioning has long been regarded as a fundamental task in visual understanding.Recently, however, few large vision-language model (LVLM) research discusses model’s image captioning performance because of the outdated short-caption benchmarks and unreliable evaluation metrics.In this work, we propose to benchmark detail image caption task by curating high-quality evaluation datasets annotated by human experts, GPT-4V and Gemini-1.5-Pro.We also design a more reliable caption evaluation metric called CAPTURE (CAPtion evaluation by exTracting and coUpling coRE information).CAPTURE extracts visual elements, e.g., objects, attributes and relations from captions, and then matches these elements through three stages, achieving the highest consistency with expert judgements over other rule-based or model-based caption metrics.The proposed benchmark and metric provide reliable evaluation for LVLM’s detailed image captioning ability.Guided by this evaluation, we further explore to unleash LVLM’s detail caption capabilities by synthesizing high-quality data through a five-stage data construction pipeline.Our pipeline only uses a given LVLM itself and other open-source tools, without any human or GPT-4V annotation in the loop.Experiments show that the proposed data construction strategy significantly improves model-generated detail caption data quality for LVLMs with leading performance, and the data quality can be further improved in a self-looping paradigm.All code and dataset will be publicly available at https://github.com/foundation-multimodal-models/CAPTURE.

footnotetext: * Equal contribution.footnotetext: †Email corresponding

1 Introduction

Image captioning has long been a fundamental task to assess LVLM’s vision understanding capability[54, 34, 12, 15].However, recent LVLM researches evaluate LVLMs’ visual understanding performance with a focus on QA benchmarks, such as MME[16], MMBench[36], MMMU[60], MM-Vet[59], etc., which may suffer from instability caused by LVLMs’ varying instruction following abilities[16].What’s worse, human-defined queries may cover a limited scope of vision features[25] and introduce bias in performance evaluation[59].Traditional image captioning task is considered unreliable for visual understanding evaluation because of the outdated benchmarks and unstable evaluation metrics.Current image caption benchmarks consist of fairly short captions with limited vision information[32, 2], while SOTA LVLMs are capable of generating detail image captions encompassing a variety of fine-grained elements[9, 54, 34], and only a few of them are covered in the provided ground truth captions.This contradiction leads to unsatisfying evaluation results.To this end, we propose to curate high-quality detail image caption evaluation datasets to provide reliable evaluation results for SOTA LVLMs.The evaluation datasets are annotated by human experts and the most capable LVLM GPT-4V[41]&Gemini-1.5-Pro[46], and are therefore of satisfying quality for state-of-the-art (SOTA) LVLM evaluation.

Apart from benchmarks, existing caption evaluation metrics also suffer from poor consistency with human judgements.Traditional rule-based caption metric such as BLEU[43], CIDER[53] and METEOR[4],compute n-gram segment matching score between candidate and reference captions, which is extremely sensitive to caption writing style, resulting into unstable evaluation results[19].Model-based evaluation metric are proposed to improve the reliability of image caption evaluation.However, representative model-based metrics either adopt outdated backbone models[3], or suffer from limited input text length[19, 48], leading to unsatisfying detail caption evaluation results.

To tackle the aforementioned problems, we propose CAPTURE, which adopts the SOTA text scene graph parser Factual[29] to extract visual elements from captions, i.e., objects, attributes and relations.We match the extracted elements from candidate and ground truth captions through a stop words filtering module and a three-stage matching strategy.Compared with SPICE, our proposed CAPTURE metric adopts a T5-based language model as parser rather than PCFG, while we design a more capable three-stage core information coupling module to match the parsed result.As illustrated in Figure1, CAPTURE produces satisfying consistency with human evaluation results, while other metrics do not.Experiments on both GPT-4 annotated dataset and human-annotated datasets show that the proposed CAPTURE achieves the highest consistency with human or GPT-4 experts, surpassing all traditional caption evaluation metrics and model-based metrics.

With CAPTURE providing reliable evaluation results, we further explore to unleash LVLMs’ detail image caption capabilities in a divide-and-conquer paradigm with a given LVLM.No expert annotation is required in our proposed data construction loop.The data construction pipeline is illustrated in Figure1.We adopt a divide-and-conquer strategy to synthesize high-quality detail image caption.An LVLM is instructed to generation both overall caption for the image and local captions for salient objects segmented by SAM[23].We adopt a novel phrase-level filtering strategy to suppress hallucinations, which extracts visual element phrases from captions, and filter out those scored low by the open-vocabulary object detection model.Finally, the filtered overall caption and local captions are fed to an LLM to be merged into a high-quality detail image caption.Experiments show that our data construction pipeline produces significantly higher-quality detail caption, and a simple-yet-effective self-looping strategy can further improve the data quality.Moreover, the synthesized data improves LVLM’s understanding capabilities effectively when incorporated into the training process.

To summarize, the contribution of this work can be listed as follows:

(1) We release a 4870-case GPT-4V&Gemini-1.5-Pro annotated detail image caption benchmark for reliable evaluation, accompanied with three model-generated captions and corresponding GPT-4 annotated quality scores for expert judgement consistency evaluation.

(2) We propose a novel detail image caption evaluation metric CAPTURE, which adopts a T5-based parser to extract visual elements from captions, and compute the matching score via a three-stage matching module.Experiments indicate that CAPTURE metric achieves the highest consistency with human and GPT-4V&Gemini-1.5-Pro judgement over other caption metrics, providing reliable detail caption evaluation results without expensive LLM API calls.

(3) We propose a five-stage detail image caption data construction pipeline, which explores to use a given LVLM and open-source vision and language tools to produce higher-quality detail caption data.Experiments show that our data construction pipeline improves detail caption data quality significantly, and the data quality can be further improved by self-looping.

2 Related Work

Image caption evaluation.

Early image captioning benchmarks, such as COCO[11], NoCaps[2], consist of precise annotated captions but contain limited visual information, which is outdated for recently released LVLMs with leading performance.Traditional caption evaluation metrics, such BLEU[43], CIDER[53] and METEOR[4], compute n-gram matching score and therefore suffer from instability caused by varying writing styles.Model-based metric SPICE[3] extracts visual elements from caption sentences, and match them to obtain evaluation results.CLIP-Score[19], MID[22] and PAC-S[48] borrow pretrained CLIP[44] model to assess the quality of model-generated image captions.Although producing relatively reliable evaluation results, these metrics can hardly tackle detail caption evaluation tasks because of the outdated backbone model (SPICE) and limited text input length (CLIPScore).

Detail caption data construction.

A series of work seek to construct detail caption data for LVLM training.ShareGPT4V[9] and ALLaVA[8] curate detail image caption data annotated by GPT-4V for model training.All-Seeing[55] leverages LLMs to imagine co-occurrence visual elements for detail caption construction.GLaMM[45] and ASMv2[56] use open-source suites for dense caption generation, with a focus on correspondence of local descriptions and image regions.Our proposed data construction pipeline adopts a divide-and-conquer strategy, unleashing LVLM’s detail caption ability by generating and merging local captions.A recent work Monkey[28] also adopts a zoom-in-and-caption approach, but they use outdated local captioner and rely on ChatGPT for caption generation.Compared with Monkey, we use open-source LVLM and LLM to synthesize detail caption data, and propose a phrase-level filtering strategy.Guided by the proposed benchmark, we also provide in-depth analysis for the effectiveness of the detail caption construction pipeline.

3 Benchmarking Detail Image Caption

In this section, we elaborate the expert judgement data construction process and the workflow of the proposed detail image caption metric.

3.1 Detail Caption Evaluation Datasets

BenchmarkData sourceAnnt. expertImg numRef numAvg lenUni. 2-gram
COCOtestsubscriptCOCO𝑡𝑒𝑠𝑡\text{COCO}_{test}COCO start_POSTSUBSCRIPT italic_t italic_e italic_s italic_t end_POSTSUBSCRIPTCOCO[32]Human500050005000500025,0102501025,01025 , 01010.5910.5910.5910.5961,4486144861,44861 , 448
NocapsvalsubscriptNocaps𝑣𝑎𝑙\text{Nocaps}_{val}Nocaps start_POSTSUBSCRIPT italic_v italic_a italic_l end_POSTSUBSCRIPTOpenimages[24]Human450045004500450045,0004500045,00045 , 00011.4911.4911.4911.49116,969116969116,969116 , 969
DetailCaps-100COCO[32], SAM[23]Human100100100100100100100100175.96175.96175.96175.9610,8581085810,85810 , 858
LAION[49], CC[50], SBU[42]
DetailCaps-4870COCO[32], SAM[23], LAION[49]GPT-4V48704870487048709740974097409740122.06122.06122.06122.06377,184377184377,184377 , 184
CC[50], SBU[42], Coyo[6], Flickr[57]Gemini-1.5-Pro

To benchmark detail image caption task reliably and better evaluate the consistency between each image caption metric and expert evaluation, we construct two expert-annotated datasets for performance evaluation.

For human evaluation dataset, we curate 100 cases sampled from ShareGPT4V-102k[9] randomly.We first call GPT-4V to generate detail captions, followed by human experts removing hallucinations and supplementing omitted visual elements.The refined detail image captions are then used as the ground truth for evaluation.We prompt three LVLMs with leading detail captioning performance for caption generation, which are ShareCaptioner[9], CogVLM[54] and LLaVA-1.5[33].Human experts are instructed to score each caption based on the precision and recall of three types of visual elements: object, attribute and relation.The overall scores range in [0,5]05[0,5][ 0 , 5 ], and are normalized to [0,1]01[0,1][ 0 , 1 ] for fair expert judgement consistency evaluation of caption metrics.

We further curate a 4870 case dataset annotated by GPT-4V&Gemini-1.5-Pro for detail caption evaluation.Besides the data sources used in human-annotated 100 cases, we further incorporate pictures from COYO[6], LAION[49], CC[7] and Flickr[58] for diversity.Captions generated by ShareCaptioner, CogVLM and LLaVA-1.5 and corresponding annotated caption scores are provided for each sample.We instruct text-only GPT-4[1] to compare model-generated captions with GPT-4V&Gemini-1.5-Pro annotated references to obtain evaluation scores.We use text-only GPT-4 for evaluation because of its outstanding instruction following abilities.We refer to AppendixA for more details about the prompts used for detail caption generation and GPT4 evaluation generation.

We show the statistics of the curated expert judgement datasets in Table1.Our detail caption evaluation benchmarks contain image samples from various sources, and the reference captions are significantly longer than previous benchmarks.It worth noticing that DetailCaps-4870 benchmark contains 377,184 unique 2-grams in 9740 reference captions, while has only 116,969 unique 2-grams across 45,000 references.

3.2 CAPTURE Metric

CAPTURE metric extracts and matches core visual elements instead of n-gram pieces to obtain evaluation results, suppressing the influence of varying writing styles.We elaborate the design of CAPTURE metric in the following parts: visual elements extraction, stop words filtering and visual elements matching.We refer to AppendixB for implementation details of CAPTURE metric.

Visual elements extraction.

Visual elements extraction module extracts objects, attributes and relations from caption sentences.We adopt Factual parser[30], which is a T5-base model with leading performance in text scene graph parsing.Since Factual parser is trained on short caption parsing dataset, we use NLTK toolkit[5] to split detail image caption into sentences to be parsed separately.The parsing results are then lemmatized (Wordnet[39]), deduplicated and merged to be the final parsing result.

Stop words filtering.

Factual parser may extract abstract nouns as object elements, for example "foreground", "background", which do not correspond to visual elements in the image, and are not expected to participate in the matching process.To this end, we curate a stop word list to filter out these abstract nouns from extracted object elements.We first apply LLaMA2-13B-chat[52] and Factual parser to ShareGPT4V-102k dataset for nouns extraction respectively, and curate words recalled by Factual parser but omitted by LLaMA2-13B-chat.We compute the frequency of these words and task human experts to judge whether words with the highest frequencies have tangible meanings.Finally, 317 words with high frequency are included in the stop word list.

Visual elements matching.

In this part, we match the extracted visual elements to produce evaluation result.We implement a three-stage matching strategy to obtain matching results, which is robust to varying writing styles.An illustration of the matching module is shown in Figure2.We first match the same visual elements, followed by a synonym matching module.Words sharing one or more synonyms are considered matched, whereWordnet[39] is employed to get the synonym set of visual elements.Phrases matched in exact or synonym matching module obtain a 1.01.01.01.0 matching score.To deal with the remaining unmatched elements,we further propose a soft matching module, which uses Sentence BERT[14] model to compute soft matching score.To be specific, we use Sentence BERT to encode the remaining object, attribute and relation phrases and compute the cosine similarity matrix between ground truth phrase embeddings and candidate ones.The max similarity score of each row and column, which is in [0.0,1.0)0.01.0[0.0,1.0)[ 0.0 , 1.0 ), are the added up to the exact matching and synonym matching scores.We then compute the precision, recall and F1 of visual elements based on the matching score.CAPTURE metric computes the caption quality score as a weighted summation of the three F1 scores, which is illustrated in Figure2.We set weights for each type of visual elements as Object:Attribute:Relation=5:5:2 by default.

4 Improving Detail Image Caption

In this section, we elaborate the design of the proposed detail caption synthesizing pipeline, and introduce how to improve LVLM training with constructed detail caption data.

4.1 Detail Caption Construction

We introduce the proposed divide-and-conquer detail caption construction pipeline in the following five stages.The pipeline is illustrated in the right part of Figure1.

Stage I: Overall caption generation.

We first instruct a given LVLM to generate overall image caption as the skeleton for high quality detail caption generation.The overall caption may suffer from hallucinations and omissions, and will be polished in the following stages.

Stage II: Salient visual elements detection.

To locate salient objects for local caption generation, we segment the image with SAM[23] and filter out masks with extreme large or small sizes.Then, we adopt a maximal rectangle algorithm to reduce overlap between remaining masks.The resulted cropped bounding boxes are regarded as salient visual elements.

Stage III: Local caption generation.

To produce complementary detail visual information for the overall caption, we instruct the given LVLM to generate local caption for each bounding box obtained in Stage II.We limit the output length of local captions to be no more than twenty words to suppress unexpected hallucinations.

Stage IV: Hallucination filtering.

We propose a novel phrase-level filtering strategy to suppress hallucinations and preserve the recalled visual elements.We first extract visual element phrases from both overall caption and local captions with Factual parser, and filter out those scored lower than 0.01 by Owlv2[40], which is an open-vocabulary object detection model.Notice that captions may suffer from some grammar errors with phrases filtered out.These errors will be corrected in the final stage.

Stage V: Caption merging.

In this stage, an LLM is instructed to merge local captions into the skeleton provided in the overall caption smoothly, instead of simply concatenating them.

With local caption providing supplementary visual information and filtering module tackling accompanied hallucinations, the synthesized detail image caption captures more visual elements with hallucinations suppressed.Visualized examples are shown in Appendix C.

4.2 Improving LVLM Training with Synthesized Detail Caption Data

We further explore to enhance LVLM’s overall understanding performance with self-generated detail caption data.We synthesize detail caption data for images from ShareGPT4V-102k dataset[9], and then select a proportion of synthesized detail caption data for model training.Samples with the largest number of visual elements extracted by Factual parser are selected for their rich visual information.The selected data is incorporated into the SFT dataset to improve overall understanding performance.

5 Experiments

In this section, we introduce the experiment settings and show main experimental results to demonstrate the effectiveness of the proposed detail image caption metric and data construction pipeline.

5.1 Benchmarking Detail Image Caption

5.1.1 Experiment Settings

Datasets.

We conduct experiments on the two expert judgement datasets described in Section3.1.Each sample in the two datasets contains expert-annotated reference detail captions, and expert-annotated caption quality scores for three SOTA LVLM-generated captions.The statistics of the two datasets are shown in Table1.

Evaluation protocol.

We evaluate the caption metrics’ consistency with expert judgements with four metrics: Pearson correlation coefficient (PCC) ρ𝜌\rhoitalic_ρ, coefficient of determination R2superscript𝑅2R^{2}italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, Kendall’s τ𝜏\tauitalic_τ (Kd τ𝜏\tauitalic_τ) and Sample τ𝜏\tauitalic_τ (Sp τ𝜏\tauitalic_τ).PCC reflects the linear correlation between the metric-evaluated scores and the expert-annotated ones.Coefficient of determination evaluates both the linear correlation and the variation of metric-evaluated score values from expert judgement.Kd τ𝜏\tauitalic_τ is computed as the proportion of matched score order pairs among all partial order pairs.Sp τ𝜏\tauitalic_τ computes Kd τ𝜏\tauitalic_τ for each sample’s caption scores independently, and use the average value as final result.Sp τ𝜏\tauitalic_τ’s formulation fits LVLM’s caption evaluation process well, and therefore is regarded as the most important metric for consistency evaluation.

Baselines.

We compare the CAPTURE metric with both rule-based and model-based caption metrics.BLEU-2[43], CIDER[53], ROUGE-L[31] and METEOR[4] are considered as representative rule-based metrics.For model-based metrics, we consider SPICE[3], CLIPScore[19] and PAC-S[48].SPICE is built on a PCFG text parser model for information extraction, while CLIPScore and PAC-S borrow CLIP model to evaluate the alignment between images and text captions.We implement the model-based metrics with OpenCLIP-L/14[21], and truncate the detail caption paragraph for alignment score computation due to the limitation in input length.We also evaluate the consistency between GPT-Eval and human judgements on DetailCaps-100 benchmark.

5.1.2 Main Results

MetricDetailCaps-100DetailCaps-4870Average
PCC ρ𝜌\rhoitalic_ρ1R21superscript𝑅21-R^{2}1 - italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPTKd τ𝜏\tauitalic_τSp τ𝜏\tauitalic_τPCC ρ𝜌\rhoitalic_ρ1R21superscript𝑅21-R^{2}1 - italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPTKd τ𝜏\tauitalic_τSp τ𝜏\tauitalic_τPCC ρ𝜌\rhoitalic_ρ \uparrow1R21superscript𝑅21-R^{2}1 - italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT \downarrowKd τ𝜏\tauitalic_τ \uparrowSp τ𝜏\tauitalic_τ \uparrow
Rule-based metrics
BLEU[2002]0.21500.21500.21500.215096.2796.2796.2796.270.16230.16230.16230.16230.21630.21630.21630.21630.30990.30990.30990.309924.8624.8624.8624.860.21350.21350.21350.21350.28120.28120.28120.28120.26250.26250.26250.262560.5760.5760.5760.570.18790.18790.18790.18790.24880.24880.24880.2488
ROUGE[2004]0.25540.25540.25540.2554185.69185.69185.69185.690.19050.19050.19050.19050.33210.33210.33210.33210.32910.32910.32910.329190.4290.4290.4290.420.23480.23480.23480.23480.33030.33030.33030.33030.29230.29230.29230.2923138.06138.06138.06138.060.21270.21270.21270.21270.33120.33120.33120.3312
METEOR[2005]0.36430.36430.36430.3643384.58384.58384.58384.580.26790.26790.26790.26790.35290.35290.35290.35290.43860.43860.43860.4386193.46193.46193.46193.460.31650.31650.31650.31650.46210.46210.46210.46210.40150.40150.40150.4015289.02289.02289.02289.020.29220.29220.29220.29220.40750.40750.40750.4075
CIDER[2015]0.08340.08340.08340.08341.7e71.7superscript𝑒71.7e^{7}1.7 italic_e start_POSTSUPERSCRIPT 7 end_POSTSUPERSCRIPT0.11590.11590.11590.11590.05640.05640.05640.05640.12130.12130.12130.12132.27e72.27superscript𝑒72.27e^{7}2.27 italic_e start_POSTSUPERSCRIPT 7 end_POSTSUPERSCRIPT0.09080.09080.09080.09080.09480.09480.09480.09480.10240.10240.10240.10241.99e71.99superscript𝑒71.99e^{7}1.99 italic_e start_POSTSUPERSCRIPT 7 end_POSTSUPERSCRIPT0.10340.10340.10340.10340.07560.07560.07560.0756
Model-based metrics
SPICE[2016]0.35800.35800.35800.3580126.60126.60126.60126.600.26410.26410.26410.26410.38190.38190.38190.38190.51550.51550.51550.5155131.1131.1131.1131.10.38180.38180.38180.38180.55540.55540.55540.55540.43680.43680.43680.4368128.85128.85128.85128.850.32300.32300.32300.32300.46870.46870.46870.4687
CLIP-Score[2021]0.25320.25320.25320.253248.8148.8148.8148.810.18070.18070.18070.18070.29280.29280.29280.29280.44630.44630.44630.446316.1016.1016.1016.100.30390.30390.30390.30390.41090.41090.41090.41090.34980.34980.34980.349832.4632.4632.4632.460.24230.24230.24230.24230.35190.35190.35190.3519
PAC-S[2023]0.25840.25840.25840.258462.8262.8262.8262.820.18330.18330.18330.18330.28430.28430.28430.28430.27830.27830.27830.278318.9318.9318.9318.930.17950.17950.17950.17950.29300.29300.29300.29300.26840.26840.26840.268440.8840.8840.8840.880.18140.18140.18140.18140.28870.28870.28870.2887
CAPTURE0.47350.4735\bm{0.4735}bold_0.473511.5811.58\bm{11.58}bold_11.580.36880.3688\bm{0.3688}bold_0.36880.61170.6117\bm{0.6117}bold_0.61170.53660.5366\bm{0.5366}bold_0.53664.824.82\bm{4.82}bold_4.820.39560.3956\bm{0.3956}bold_0.39560.57370.5737\bm{0.5737}bold_0.57370.50510.5051\bm{0.5051}bold_0.50518.208.20\bm{8.20}bold_8.200.38220.3822\bm{0.3822}bold_0.38220.59270.5927\bm{0.5927}bold_0.5927
GPT4-Eval0.515744.440.42370.6120----------------

CAPTURE achieves the highest consistency with expert judgements.

As shown in Table2, the proposed metric CAPTURE improves PCC ρ𝜌\rhoitalic_ρ by 0.06830.06830.06830.0683 (15.6%percent15.6absent15.6\%\uparrow15.6 % ↑), R2superscript𝑅2R^{2}italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT score by 24.2624.2624.2624.26 (74.7%percent74.7absent74.7\%\downarrow74.7 % ↓), Kd τ𝜏\tauitalic_τ by 0.05920.05920.05920.0592 (18.3%percent18.3absent18.3\%\uparrow18.3 % ↑) and Sp τ𝜏\tauitalic_τ by 0.12400.12400.12400.1240 (26.4%percent26.4absent26.4\%\uparrow26.4 % ↑) over previous SOTA baselines.The advantages in PCC ρ𝜌\rhoitalic_ρ, Kd τ𝜏\tauitalic_τ and Sp τ𝜏\tauitalic_τ indicate that the proposed metric performs the best in linear correlation with expert judgment and pair-wise ranking accuracy, showing promising prospects for LVLM-generated detail caption evaluation.Besides, CAPTURE also performs the best in 1R21superscript𝑅21-R^{2}1 - italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT metric, indicating that CAPTURE produces evaluation results with aligned values.

METEOR and SPICE perform the best among rule-based and model-based metrics, respectively.

We attribute METEOR’s satisfying performance to its consideration for both precision and recall of n-grams.METEOR also adopts exact, synonym and porter stem matching strategies, improving its robustness to varying writing styles.For SPICE, its PCFG parser performs more robust for long detail captions compared with CLIP-based metrics, which suffer from CLIP’s limited input text length.

GPT4-Eval achieves the highest consistency with human evaluation on DetailCaps-100 dataset.

This result validates the effectiveness of evaluating caption metrics’ consistency with GPT4-Eval results on the larger dataset DetailCaps-4870.It is also worth noticing that CATURE’s consistency performance is pretty close to that of GPT-Eval.Moreover, CAPTURE does not require calling expensive LLM APIs, demonstrating its promising prospect in detail caption evaluation.

5.1.3 Analysis

We verify the effectiveness of the design of CAPTURE metric.Among the consistency evaluation metrics, we point out that Sp τ𝜏\tauitalic_τ is the closest to real detail caption evaluation scenario, and we focus on this metrics for analysis.

Stop words filtering improves sample-level evaluation consistency effectively.

Statistics show that when evaluating candidate captions on DetailCaps-100 dataset, 28.43% extracted object phrases are detected and discarded by the stop words filtering module.As shown in Table3, performance drops on SP τ𝜏\tauitalic_τ are witnessed on both DetailCaps-100 and DetailCaps-4870 benchmark when stop words filtering module is removed.We attribute the fluctuation in other consistency metrics to the varying number of visual elements discarded by the stop words filtering module across samples.

Soft matching module improves evaluation consistency and the alignment of evaluation score values.

When soft matching module is removed, CAPTURE suffers from a 3.3%percent3.3absent3.3\%\downarrow3.3 % ↓ performance drop in Sp τ𝜏\tauitalic_τ.It is also worth noticing that 1R21superscript𝑅21-R^{2}1 - italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT score deteriorates the most significantly.The soft matching strategy tackles a variety of phrases with similar meaning, and thus makes up the deficiency of exact matching and synonym matching modules when tackling varying writing styles.

The default α,β,γ=5,5,2formulae-sequence𝛼𝛽𝛾552\alpha,\beta,\gamma=5,5,2italic_α , italic_β , italic_γ = 5 , 5 , 2 setting is a sweet spot for detail caption evaluation.

We modify the scale factors of relation elements γ𝛾\gammaitalic_γ from 00 (discarding relation matching score) to 5555 (relation F1 is considered equally with object F1 and attribute F1) to verify this judgement.Experiment results show that CAPTURE’s performance drops with relation matching score ratio γ𝛾\gammaitalic_γ as 00 or 5555, validating that α,β,γ=5,5,2formulae-sequence𝛼𝛽𝛾552\alpha,\beta,\gamma=5,5,2italic_α , italic_β , italic_γ = 5 , 5 , 2 is the most suitable for CAPTURE’s evaluation.

MetricDetailCaps-100DetailCaps-4870Average
PCC ρ𝜌\rhoitalic_ρ1R21superscript𝑅21-R^{2}1 - italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPTKd τ𝜏\tauitalic_τSp τ𝜏\tauitalic_τPCC ρ𝜌\rhoitalic_ρ1R21superscript𝑅21-R^{2}1 - italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPTKd τ𝜏\tauitalic_τSp τ𝜏\tauitalic_τPCC ρ𝜌\rhoitalic_ρ \uparrow1R21superscript𝑅21-R^{2}1 - italic_R start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT \downarrowKd τ𝜏\tauitalic_τ \uparrowSp τ𝜏\tauitalic_τ \uparrow
CAPTURE0.47350.47350.47350.473511.5811.5811.5811.580.36880.36880.36880.36880.61170.61170.61170.61170.53660.53660.53660.53664.824.824.824.820.39560.39560.39560.39560.57370.57370.57370.57370.50510.50510.50510.50518.208.208.208.200.38220.38220.38220.38220.59270.59270.59270.5927
- Stop words0.48300.48300.48300.483013.2313.2313.2313.230.38040.38040.38040.38040.59470.59470.59470.59470.53560.53560.53560.53565.945.945.945.940.39400.39400.39400.39400.57300.57300.57300.57300.50930.50930.50930.50939.589.589.589.580.38720.38720.38720.38720.58380.58380.58380.5838
- Soft matching0.46740.46740.46740.467429.1529.1529.1529.150.34880.34880.34880.34880.57700.57700.57700.57700.54260.54260.54260.542618.6518.6518.6518.650.39640.39640.39640.39640.56850.56850.56850.56850.50500.50500.50500.505023.9023.9023.9023.900.37260.37260.37260.37260.57280.57280.57280.5728
α,β,γ=5,5,0formulae-sequence𝛼𝛽𝛾550\alpha,\beta,\gamma=5,5,0italic_α , italic_β , italic_γ = 5 , 5 , 00.46540.46540.46540.46549.219.219.219.210.36420.36420.36420.36420.59470.59470.59470.59470.52820.52820.52820.52823.943.943.943.940.39410.39410.39410.39410.56860.56860.56860.56860.49680.49680.49680.49686.586.586.586.580.37920.37920.37920.37920.58160.58160.58160.5816
α,β,γ=5,5,5formulae-sequence𝛼𝛽𝛾555\alpha,\beta,\gamma=5,5,5italic_α , italic_β , italic_γ = 5 , 5 , 50.46510.46510.46510.465113.7513.7513.7513.750.35560.35560.35560.35560.60640.60640.60640.60640.52700.52700.52700.52705.655.655.655.650.38340.38340.38340.38340.56590.56590.56590.56590.49600.49600.49600.49609.709.709.709.700.36950.36950.36950.36950.58620.58620.58620.5862

LVLMLanguageDetail Caption DataResolutionDC100subscriptDC100\text{DC}_{100}DC start_POSTSUBSCRIPT 100 end_POSTSUBSCRIPTDC4870subscriptDC4870\text{DC}_{4870}DC start_POSTSUBSCRIPT 4870 end_POSTSUBSCRIPT
CogVLM[2023a]Vicuna-7BHuman Annt.4902superscript4902490^{2}490 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT63.0163.0163.0163.0160.2060.2060.2060.20
ShareCaptioner-7B[2023a]Vicuna-7BGPT-4V Annt.4482superscript4482448^{2}448 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT60.8560.8560.8560.8559.5559.5559.5559.55
LLaVA-1.5-7B[2023a]Vicuna-7BSynthesized3362superscript3362336^{2}336 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT51.2351.2351.2351.2349.6949.6949.6949.69
LLaVA-1.5-13B[2023a]Vicuna-13BSynthesized3362superscript3362336^{2}336 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT51.7451.7451.7451.7451.1051.1051.1051.10
LLaVA-NEXT-7B[2024a]Vicuna-7BGPT-4V Annt.3362superscript3362336^{2}336 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT*{1111-5555}60.1860.1860.1860.1858.2258.2258.2258.22
LLaVA-NEXT-13B[2024a]Vicuna-13BGPT-4V Annt.3362superscript3362336^{2}336 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT*{1111-5555}60.3860.3860.3860.3858.6658.6658.6658.66
LLaVA-NEXT-34B[2024a]Hermes-2-Yi-34BGPT-4V Annt.3362superscript3362336^{2}336 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT*{1111-5555}60.6060.6060.6060.6058.8858.8858.8858.88
Mini-Gemini-HD-7B[2024]Vicuna-7BGPT-4V Annt.3362superscript3362336^{2}336 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT*555559.5159.5159.5159.5157.5857.5857.5857.58
Mini-Gemini-HD-13B[2024]Vicuna-13BGPT-4V Annt.3362superscript3362336^{2}336 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT*555560.5160.5160.5160.5158.3958.3958.3958.39
Intern-XComposerV2[2024]Vicuna-7BGPT-4V Annt.4902superscript4902490^{2}490 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT61.4361.4361.4361.4359.9259.9259.9259.92
InternVL-V1.2-PLUS-40B[2023b]Hermes-2-Yi-34BGPT-4V Annt.4482superscript4482448^{2}448 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT61.6161.6161.6161.6160.7560.7560.7560.75
InternVL-V1.5-26B[2024c]InternLM-20BGPT-4V Annt.4482superscript4482448^{2}448 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT*{1111-41414141}65.6265.6265.6265.6262.3662.3662.3662.36

5.1.4 Evaluating LVLMs with Leading Performance

With DetailCaps benchmark and CAPTURE evaluating LVLMs’ detail captioning performance reliably, we review the detail caption capabilities for 12 open source LVLMs with leading performance.The evaluation results on DetailCaps-100 and DetailCaps-4870 are shown in Table4.Among all models, InternVL-V1.5[13] achieves the best detail image caption performance with a large advantage over other models.It also can be observed from the results of the LLaVA-1.5, LLaVA-Next and Mini-Gemini[26] series that model’s detail captioning ability improves consistently as the model size increases.In addition, a common observation is that training with detail caption data generated by GPT-4V leads to better detail captioning performance.Among these LVLMs, CogVLM achieves the second highest CAPTURE score with high-quality human-refined detail image caption data.

5.2 Improving Detail Image Caption

5.2.1 Experiment Settings

We use ShareGPT4V-102k dataset for detail caption data construction and implement two pipelines with different model size.For 7B model pipeline, we use SAM-ViT-L[23] for segmentation, LLaVA-1.5-7B for overall and local caption generation, OwlV2-large-ensemble[40] for hallucination filtering and LLaMA-2-7B-Chat for caption mering.For 13B model pipeline, we replace SAM-ViT-H, LLaVA-1.5-13B, and LLaMA-2-13B-Chat instead.We validate the effectiveness of the proposed data construction pipeline with four LVLMs with leading performance, which are LLaVA-1.5-7B, LLaVA-1.5-13B, LLaVA-NEXT-7B and Mini-Gemini-7B-HD.

CaptionDetailcaps-100Detailcaps-4870Average
CAPTUREPrecisionRecallCAPTUREPrecisionRecallCAPTUREPrecisionRecall
LLaVA-1.5-7B
Self51.2351.2351.2351.2365.2465.2465.2465.2443.3143.3143.3143.3151.2751.2751.2751.2764.6164.6164.6164.6143.7943.7943.7943.7951.2551.2551.2551.2564.9264.9264.9264.9243.5543.5543.5543.55
Synthesized57.1157.1157.1157.1164.1264.1264.1264.1252.0852.0852.0852.0856.1856.1856.1856.1863.0263.0263.0263.0251.4851.4851.4851.4856.6456.6456.6456.6463.5763.5763.5763.5751.7851.7851.7851.78
LLaVA-1.5-13B
Self51.7651.7651.7651.7665.0165.0165.0165.0144.1044.1044.1044.1051.4551.4551.4551.4565.0465.0465.0465.0443.9143.9143.9143.9151.6151.6151.6151.6165.0365.0365.0365.0344.0044.0044.0044.00
Synthesized57.3657.3657.3657.3662.0762.0762.0762.0753.5253.5253.5253.5256.8356.8356.8356.8361.6161.6161.6161.6153.353.353.353.357.0957.0957.0957.0961.8461.8461.8461.8453.4153.4153.4153.41
LLaVA-NEXT-7B
Self61.4861.4861.4861.4865.6065.6065.6065.6057.8257.8257.8257.8259.8659.8659.8659.8664.1664.1664.1664.1656.6156.6156.6156.6160.6760.6760.6760.6764.8864.8864.8864.8857.2257.2257.2257.22
Synthesized62.2462.2462.2462.2464.4964.4964.4964.4960.0760.0760.0760.0760.1060.1060.1060.1062.3662.3662.3662.3658.6058.6058.6058.6061.1761.1761.1761.1763.4263.4263.4263.4259.3459.3459.3459.34
Mini-Gemini-7B-HD
Self59.5159.5159.5159.5161.9961.9961.9961.9957.2857.2857.2857.2857.6857.6857.6857.6860.2460.2460.2460.2455.8955.8955.8955.8958.5958.5958.5958.5961.1261.1261.1261.1256.5956.5956.5956.59
Synthesized60.4460.4460.4460.4460.9860.9860.9860.9859.7859.7859.7859.7858.6458.6458.6458.6458.7658.7658.7658.7659.1759.1759.1759.1759.5459.5459.5459.5459.8759.8759.8759.8759.4859.4859.4859.48

5.2.2 Main results

Our detail caption synthesizing pipeline improves LVLM-generated caption quality effectively.

As shown in Table5, for LLaVA-1.5-7B and LLaVA-1.5-13B, the detail caption quality is improved by a large fraction in terms of CAPTURE score.For more advanced LVLM like LLaVA-NEXT and Mini-Gemini-HD, the advantage of the proposed pipeline persists, demonstrating the effectiveness of the our data synthesizing strategy.We attribute the smaller fraction of improvement in LLaVA-NEXT and Mini-Gemini-HD to other vision and language tools’ limited capabilities, which pose "short boards" compared with LVLMs trained with expert-annotated detail caption training data.

Our pipeline enhances recall of visual elements effectively with little precision drop.

As shown in Table5, this tendency can be observed across all four LVLMs, indicating that the divide-and-conquer strategy improves model’s perception of detail visual elements effectively.Thanks to the hallucination filtering module, the performance drop in precision is suppressed, so that improvement on CAPTURE score is witnessed across all LVLMs.

5.2.3 Analysis

CaptionDetailcaps-100Detailcaps-4870Average
CAPTUREPrecisionRecallCAPTUREPrecisionRecallCAPTUREPrecisionRecall
Ablation
Self51.2351.2351.2351.2365.2465.2465.2465.2443.3143.3143.3143.3151.2751.2751.2751.2764.6164.6164.6164.6143.7943.7943.7943.7951.2551.2551.2551.2564.9264.9264.9264.9243.5543.5543.5543.55
Synthesized57.1157.1157.1157.1166.3166.3166.3166.3152.1652.1652.1652.1656.1856.1856.1856.1863.0263.0263.0263.0251.4851.4851.4851.4856.6456.6456.6456.6464.6764.6764.6764.6751.8251.8251.8251.82
- filter56.7856.7856.7856.7865.1665.1665.1665.1653.2653.2653.2653.2656.0156.0156.0156.0162.8062.8062.8062.8051.5151.5151.5151.5156.3956.3956.3956.3963.9863.9863.9863.9852.3852.3852.3852.38
vqa filter56.4456.4456.4456.4463.9563.9563.9563.9551.1151.1151.1151.1155.8555.8555.8555.8562.8062.8062.8062.8051.1251.1251.1251.1256.1456.1456.1456.1463.3863.3863.3863.3851.1151.1151.1151.11
filter local56.7556.7556.7556.7563.8763.8763.8763.8751.6151.6151.6151.6156.2456.2456.2456.2462.7462.7462.7462.7451.9051.9051.9051.9056.5056.5056.5056.5063.3063.3063.3063.3051.7551.7551.7551.75
Self-looping
Self51.2351.2351.2351.2365.2465.2465.2465.2443.3143.3143.3143.3151.2751.2751.2751.2764.6164.6164.6164.6143.7943.7943.7943.7951.2551.2551.2551.2564.9264.9264.9264.9243.5543.5543.5543.55
loop151.9151.9151.9151.9163.4863.4863.4863.4845.0245.0245.0245.0252.5252.5252.5252.5263.7663.7663.7663.7645.8145.8145.8145.8152.2252.2252.2252.2263.6263.6263.6263.6245.4245.4245.4245.42
loop252.5052.5052.5052.5063.4363.4363.4363.4345.6645.6645.6645.6652.5352.5352.5352.5362.5762.5762.5762.5746.4246.4246.4246.4252.5252.5252.5252.5263.0063.0063.0063.0046.0446.0446.0446.04
loop352.8952.8952.8952.8962.4562.4562.4562.4546.8646.8646.8646.8652.8652.8652.8652.8661.8361.8361.8361.8347.2547.2547.2547.2552.8852.8852.8852.8862.1462.1462.1462.1447.0547.0547.0547.05
loop454.0254.0254.0254.0262.2462.2462.2462.2448.4548.4548.4548.4554.3854.3854.3854.3861.6861.6861.6861.6849.5049.5049.5049.5054.2054.2054.2054.2061.9661.9661.9661.9648.9848.9848.9848.98

Our phrase-level hallucination filtering strategy achieves the best performance.

As shown in Table6, when the filtering module is removed (-filter), a performance drop in CAPTURE score is witnessed.We also compare our filtering strategy with other alternatives used in Monkey[28].For VQA filtering, we use LVLM to check if the visual element phrase exists in the image.For local caption filtering, we filter out hallucinated local caption sentences rather than extracted phrases.Experiment results show that both alternatives lead to performance drops in CAPTURE score, demonstrating the effectiveness of the proposed phrase-level filtering strategy.

LVLM’s detail caption ablity can be improved via self-looping.

We adopt LLaVA-1.5-7B as the backbone LVLM, and synthesize detail caption data for model’s training.In each loop, we rerun the SFT stage of LLaVA-1.5-7B from a pretrain checkpoint (without any SFT), with annotated 25k detail caption data incorporated into the training data.Experiment results are shown in Table6.Model’s detail captioning ability keeps improving in the listed 4 loops, showing a promising self-evolutioning phenomena in detail captioning performance.

5.2.4 Improving LVLM Training with Synthesized Detail Caption Data

Experiment Settings.

We follow LLaVA-1.5[33] pipeline for model training.The vision-language projector is trained with 558k short caption data and a 128 batch size during pretraining, and all parameters except the vision module are trained with 665k visual instruction tuning data and a 256 batch size during SFT.We train the model with AdamW optmizer, with a 1e41superscript𝑒41e^{-4}1 italic_e start_POSTSUPERSCRIPT - 4 end_POSTSUPERSCRIPT pretraining learning rate and a 2e52superscript𝑒52e^{-5}2 italic_e start_POSTSUPERSCRIPT - 5 end_POSTSUPERSCRIPT SFT learning rate.We add 25k detail caption data into SFT stage for the 7B model, and 50k for the 13B model due to its larger capacity.In our experiments, the pretraining process takes 24 GPU hours and SFT takes 88 GPU hours on Nvidia A100.We use MME[17], MMMU[60], MMStar[10], GQA[20], VizWiz[18], POPE[27] and the proposed DetailCaps benchmarks for model’s natural scene visual understanding ability evaluation.RefCOCOg[38] is a referring expression comprehension task to evaluate model’s detail understanding capability.OCRBench[37] and DocVQA[51] are selected to evaluate model’s performance in text-heavy scenarios.For baselines, we report our reproduced results rather than reported ones for fair comparison.

DC DataMMEpsubscriptMMEp\text{MME}_{\text{p}}MME start_POSTSUBSCRIPT p end_POSTSUBSCRIPTMMEcsubscriptMMEc\text{MME}_{\text{c}}MME start_POSTSUBSCRIPT c end_POSTSUBSCRIPTMMMUvsubscriptMMMUv\text{MMMU}_{\text{v}}MMMU start_POSTSUBSCRIPT v end_POSTSUBSCRIPTMMStarGQAVisWizPOPERefCOCOgsubscriptRefCOCOg\text{RefCOCO}_{\text{g}}RefCOCO start_POSTSUBSCRIPT g end_POSTSUBSCRIPTOCRBenchsubscriptOCRBench\text{OCR}_{\text{Bench}}OCR start_POSTSUBSCRIPT Bench end_POSTSUBSCRIPTVQADocsubscriptVQADoc\text{VQA}_{\text{Doc}}VQA start_POSTSUBSCRIPT Doc end_POSTSUBSCRIPTDC100subscriptDC100\text{DC}_{\text{100}}DC start_POSTSUBSCRIPT 100 end_POSTSUBSCRIPTDC4870subscriptDC4870\text{DC}_{\text{4870}}DC start_POSTSUBSCRIPT 4870 end_POSTSUBSCRIPTWin
LLaVA-1.5-7B
Base1487.11487.11487.11487.1260.4260.4\bm{260.4}bold_260.434.634.634.634.633.3333.3333.3333.3362.8662.8662.8662.8653.7053.7053.7053.7086.2286.2286.2286.2272.1672.1672.1672.1631631631631628.7528.7528.7528.7551.2651.2651.2651.2651.4551.4551.4551.45--
+ Self 25k1499.31499.31499.31499.3258.6258.6258.6258.636.336.336.336.333.4033.4033.4033.4062.6462.6462.6462.6454.9054.9054.9054.9086.8486.8486.8486.8472.7572.75\bm{72.75}bold_72.7531631631631629.2629.2629.2629.2651.4951.4951.4951.4951.8351.8351.8351.8310/12101210/1210 / 12
+ Syn 25k1523.21523.2\bm{1523.2}bold_1523.2257.1257.1257.1257.137.337.3\bm{37.3}bold_37.333.5333.53\bm{33.53}bold_33.5362.8662.86\bm{62.86}bold_62.8656.8856.88\bm{56.88}bold_56.8887.0887.08\bm{87.08}bold_87.0872.6172.6172.6172.61𝟑𝟐𝟏321\bm{321}bold_32130.0130.01\bm{30.01}bold_30.0151.9151.91\bm{51.91}bold_51.9152.6552.65\bm{52.65}bold_52.65𝟏𝟏/𝟏𝟐1112\bm{11/12}bold_11 bold_/ bold_12
LLaVA-1.5-13B
Base1553.41553.41553.41553.4267.1267.1267.1267.134.334.334.334.334.8034.8034.8034.8063.3663.3663.3663.3658.3558.3558.3558.3585.9085.9085.9085.9074.5174.5174.5174.5133133133133130.6030.6030.6030.6051.9651.9651.9651.9652.0552.0552.0552.05--
+ Self 50k1543.41543.41543.41543.4286.8286.8286.8286.834.334.334.334.335.4035.40\bm{35.40}bold_35.4063.5363.5363.5363.5359.0859.08\bm{59.08}bold_59.0886.1786.1786.1786.1774.6374.6374.6374.6333133133133130.6630.6630.6630.6652.6252.62\bm{52.62}bold_52.6252.8552.8552.8552.8511/12111211/1211 / 12
+ Syn 50k1564.01564.0\bm{1564.0}bold_1564.0286.8286.8\bm{286.8}bold_286.834.334.3\bm{34.3}bold_34.335.2735.2735.2735.2763.5663.56\bm{63.56}bold_63.5658.5658.5658.5658.5686.2886.28\bm{86.28}bold_86.2874.6574.65\bm{74.65}bold_74.65𝟑𝟑𝟑333\bm{333}bold_33330.7930.79\bm{30.79}bold_30.7952.5652.5652.5652.5652.9152.91\bm{52.91}bold_52.91𝟏𝟐/𝟏𝟐1212\bm{12/12}bold_12 bold_/ bold_12

Synthesized detail caption data improves LVLM’s overall understanding performance effectively.

As shown in Table7, even if we only add a little fraction of synthesized high-quality detail caption data in the SFT stage (25k for 7B model and 50k for 13B model), performance improvement is witnessed across a series of visual understanding benchmarks, demonstrating the effectiveness enhancing LVLM’s overall understanding capabilities with synthesized detail caption data.

Directly generated detail caption data also improves LVLM’s overall understanding performance.

As shown in Table7, training with detail caption data generated directly also leads to an overall performance improvement.Although the improvement is eclipsed by synthesized detail caption data, this observation validates the importance of using detail caption data for model training, even if the data is generated by the model itself directly.

Model’s benchmark scores correlate to their detail caption task performance positively.

We observe a positive correlation between LVLMs’ benchmark scores (win rates) and their performance in detail caption tasks.This observation validates the importance of detail image captioning task and the feasibility of enhancing LVLM’s overall visual understanding abilities by improving its detail caption ability with synthesized high-quality caption data.

6 Limitations and Future Work

The proposed detail image caption evaluation metric achieves outstanding consistency with human evaluation in the curated benchmarks.However, we point out that although two powerful expert are adopted for evaluation dataset construction, these captions may not be perfect.Human refining and more reference captions will be incorporated into the detail caption benchmark in out future work.For the data construction pipeline, we observe a diminishing effect when the backbone LVLM becomes stronger.For example, LVLMs like LLaVA-NEXT and Mini-Gemini uses GPT-4V-annotated detail caption data for training, and therefore the advantage of the proposed pipeline may suffer from incompatible capabilities of other vision and language tools used in the pipeline.We will seek to further improve LVLM’s detail captioning abilities with more powerful and scalable vision and language suites in our future work.

7 Conclusions

In this work, we analyze the shortcomings of existing image caption benchmarks for LVLM evaluation, and curate high-quality expert-annotated evaluation dataset for detail caption evaluation.We also propose a novel detail image caption metric CAPTURE, which extracts visual elements from detail captions, and match them through three stages to produce evaluation results.Experiments show that CAPTURE metric achieves the highest consistency with expert judgements, and ablation studies demonstrate the effectiveness of the stop words filtering module, three-stage matching module and the default ratio of different type of visual elements.Guided by the proposed detail caption evaluation methods, we further seek to unleash LVLM’s detail image captioning ability with a divide-and-conquer caption construction pipeline powered by open-source vision and language tools.Experiments show that the proposed pipeline improves LVLM-annotated detail caption data quality significantly, and the data quality can be further improved via self-looping.Ablation studies validate the effectiveness of the pipeline design.

References

  • Achiam etal. [2023]Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, FlorenciaLeoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, etal.Gpt-4 technical report.arXiv preprint arXiv:2303.08774, 2023.
  • Agrawal etal. [2019]Harsh Agrawal, Karan Desai, Yufei Wang, Xinlei Chen, Rishabh Jain, Mark Johnson, Dhruv Batra, Devi Parikh, Stefan Lee, and Peter Anderson.Nocaps: Novel object captioning at scale.In Proceedings of the IEEE/CVF international conference on computer vision, pages 8948–8957, 2019.
  • Anderson etal. [2016]Peter Anderson, Basura Fernando, Mark Johnson, and Stephen Gould.Spice: Semantic propositional image caption evaluation.In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part V 14, pages 382–398. Springer, 2016.
  • Banerjee and Lavie [2005]Satanjeev Banerjee and Alon Lavie.Meteor: An automatic metric for mt evaluation with improved correlation with human judgments.In Proceedings of the acl workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization, pages 65–72, 2005.
  • Bird [2006]Steven Bird.Nltk: the natural language toolkit.In Proceedings of the COLING/ACL 2006 Interactive Presentation Sessions, pages 69–72, 2006.
  • Byeon etal. [2022]Minwoo Byeon, Beomhee Park, Haecheon Kim, Sungjun Lee, Woonhyuk Baek, and Saehoon Kim.Coyo-700m: Image-text pair dataset.https://github.com/kakaobrain/coyo-dataset, 2022.
  • Changpinyo etal. [2021]Soravit Changpinyo, Piyush Sharma, Nan Ding, and Radu Soricut.Conceptual 12m: Pushing web-scale image-text pre-training to recognize long-tail visual concepts.In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 3558–3568, 2021.
  • Chen etal. [2024a]GuimingHardy Chen, Shunian Chen, Ruifei Zhang, Junying Chen, Xiangbo Wu, Zhiyi Zhang, Zhihong Chen, Jianquan Li, Xiang Wan, and Benyou Wang.Allava: Harnessing gpt4v-synthesized data for a lite vision-language model.arXiv preprint arXiv:2402.11684, 2024a.
  • Chen etal. [2023a]Lin Chen, Jisong Li, Xiaoyi Dong, Pan Zhang, Conghui He, Jiaqi Wang, Feng Zhao, and Dahua Lin.Sharegpt4v: Improving large multi-modal models with better captions.arXiv preprint arXiv:2311.12793, 2023a.
  • Chen etal. [2024b]Lin Chen, Jinsong Li, Xiaoyi Dong, Pan Zhang, Yuhang Zang, Zehui Chen, Haodong Duan, Jiaqi Wang, YuQiao, Dahua Lin, etal.Are we on the right way for evaluating large vision-language models?arXiv preprint arXiv:2403.20330, 2024b.
  • Chen etal. [2015]Xinlei Chen, Hao Fang, Tsung-Yi Lin, Ramakrishna Vedantam, Saurabh Gupta, Piotr Dollár, and CLawrence Zitnick.Microsoft coco captions: Data collection and evaluation server.arXiv preprint arXiv:1504.00325, 2015.
  • Chen etal. [2023b]Zhe Chen, Jiannan Wu, Wenhai Wang, Weijie Su, Guo Chen, Sen Xing, Muyan Zhong, Qinglong Zhang, Xizhou Zhu, Lewei Lu, Bin Li, Ping Luo, Tong Lu, YuQiao, and Jifeng Dai.Internvl: Scaling up vision foundation models and aligning for generic visual-linguistic tasks.arXiv preprint arXiv:2312.14238, 2023b.
  • Chen etal. [2024c]Zhe Chen, Weiyun Wang, Hao Tian, Shenglong Ye, Zhangwei Gao, Erfei Cui, Wenwen Tong, Kongzhi Hu, Jiapeng Luo, Zheng Ma, etal.How far are we to gpt-4v? closing the gap to commercial multimodal models with open-source suites.arXiv preprint arXiv:2404.16821, 2024c.
  • Devlin etal. [2018]Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova.Bert: Pre-training of deep bidirectional transformers for language understanding.arXiv preprint arXiv:1810.04805, 2018.
  • Dong etal. [2024]Xiaoyi Dong, Pan Zhang, Yuhang Zang, Yuhang Cao, Bin Wang, Linke Ouyang, Xilin Wei, Songyang Zhang, Haodong Duan, Maosong Cao, etal.Internlm-xcomposer2: Mastering free-form text-image composition and comprehension in vision-language large model.arXiv preprint arXiv:2401.16420, 2024.
  • Fu etal. [2023a]Chaoyou Fu, Peixian Chen, Yunhang Shen, Yulei Qin, Mengdan Zhang, XuLin, Zhenyu Qiu, Wei Lin, Jinrui Yang, Xiawu Zheng, KeLi, Xing Sun, and Rongrong Ji.Mme: A comprehensive evaluation benchmark for multimodal large language models.ArXiv, abs/2306.13394, 2023a.URL https://api.semanticscholar.org/CorpusID:259243928.
  • Fu etal. [2023b]Chaoyou Fu, Peixian Chen, Yunhang Shen, Yulei Qin, Mengdan Zhang, XuLin, Jinrui Yang, Xiawu Zheng, KeLi, Xing Sun, Yunsheng Wu, and Rongrong Ji.Mme: A comprehensive evaluation benchmark for multimodal large language models.arXiv preprint arXiv:2306.13394, 2023b.
  • Gurari etal. [2018]Danna Gurari, Qing Li, AbigaleJ Stangl, Anhong Guo, Chi Lin, Kristen Grauman, Jiebo Luo, and JeffreyP Bigham.Vizwiz grand challenge: Answering visual questions from blind people.In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3608–3617, 2018.
  • Hessel etal. [2021]Jack Hessel, Ari Holtzman, Maxwell Forbes, Ronan LeBras, and Yejin Choi.CLIPScore: A reference-free evaluation metric for image captioning.In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih, editors, Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 7514–7528, 2021.
  • Hudson and Manning [2019]DrewA Hudson and ChristopherD Manning.Gqa: A new dataset for real-world visual reasoning and compositional question answering.In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 6700–6709, 2019.
  • Karkkainen and Joo [2021]Kimmo Karkkainen and Jungseock Joo.Fairface: Face attribute dataset for balanced race, gender, and age for bias measurement and mitigation.In Proceedings of the IEEE/CVF winter conference on applications of computer vision, pages 1548–1558, 2021.
  • Kim etal. [2022]Jin-Hwa Kim, Yunji Kim, Jiyoung Lee, KangMin Yoo, and Sang-Woo Lee.Mutual information divergence: A unified metric for multimodal generative models.Advances in Neural Information Processing Systems, 35:35072–35086, 2022.
  • Kirillov etal. [2023]Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, AlexanderC Berg, Wan-Yen Lo, etal.Segment anything.In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4015–4026, 2023.
  • Krasin etal. [2017]Ivan Krasin, Tom Duerig, Neil Alldrin, Vittorio Ferrari, Sami Abu-El-Haija, Alina Kuznetsova, Hassan Rom, Jasper Uijlings, Stefan Popov, Andreas Veit, etal.Openimages: A public dataset for large-scale multi-label and multi-class image classification.Dataset available from https://github. com/openimages, 2(3):18, 2017.
  • Li etal. [2023a]Bohao Li, Rui Wang, Guangzhi Wang, Yuying Ge, Yixiao Ge, and Ying Shan.Seed-bench: Benchmarking multimodal llms with generative comprehension.ArXiv, abs/2307.16125, 2023a.URL https://api.semanticscholar.org/CorpusID:260334888.
  • Li etal. [2024]Yanwei Li, Yuechen Zhang, Chengyao Wang, Zhisheng Zhong, Yixin Chen, Ruihang Chu, Shaoteng Liu, and Jiaya Jia.Mini-gemini: Mining the potential of multi-modality vision language models.arXiv preprint arXiv:2403.18814, 2024.
  • Li etal. [2023b]Yifan Li, Yifan Du, Kun Zhou, Jinpeng Wang, WayneXin Zhao, and Ji-Rong Wen.Evaluating object hallucination in large vision-language models.arXiv preprint arXiv:2305.10355, 2023b.
  • Li etal. [2023c]Zhang Li, Biao Yang, Qiang Liu, Zhiyin Ma, Shuo Zhang, Jingxu Yang, Yabo Sun, Yuliang Liu, and Xiang Bai.Monkey: Image resolution and text label are important things for large multi-modal models.arXiv preprint arXiv:2311.06607, 2023c.
  • Li etal. [2023d]Zhuang Li, Yuyang Chai, TerryZhuo Yue, Lizhen Qu, Gholamreza Haffari, Fei Li, Donghong Ji, and QuanHung Tran.Factual: A benchmark for faithful and consistent textual scene graph parsing.arXiv preprint arXiv:2305.17497, 2023d.
  • Li etal. [2023e]Zhuang Li, Yuyang Chai, TerryYue Zhuo, Lizhen Qu, Gholamreza Haffari, Fei Li, Donghong Ji, and QuanHung Tran.FACTUAL: A benchmark for faithful and consistent textual scene graph parsing.In Findings of the Association for Computational Linguistics: ACL 2023, pages 6377–6390, Toronto, Canada, July 2023e. Association for Computational Linguistics.URL https://aclanthology.org/2023.findings-acl.398.
  • Lin [2004]Chin-Yew Lin.Rouge: A package for automatic evaluation of summaries.In Text summarization branches out, pages 74–81, 2004.
  • Lin etal. [2014]Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and CLawrence Zitnick.Microsoft coco: Common objects in context.In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pages 740–755. Springer, 2014.
  • Liu etal. [2023a]Haotian Liu, Chunyuan Li, Yuheng Li, and YongJae Lee.Improved baselines with visual instruction tuning.In NeurIPS 2023 Workshop on Instruction Tuning and Instruction Following, 2023a.
  • Liu etal. [2024a]Haotian Liu, Chunyuan Li, Yuheng Li, BoLi, Yuanhan Zhang, Sheng Shen, and YongJae Lee.Llava-next: Improved reasoning, ocr, and world knowledge, January 2024a.URL https://llava-vl.github.io/blog/2024-01-30-llava-next/.
  • Liu etal. [2024b]Haotian Liu, Chunyuan Li, Qingyang Wu, and YongJae Lee.Visual instruction tuning.Advances in neural information processing systems, 36, 2024b.
  • Liu etal. [2023b]Yuanzhan Liu, Haodong Duan, Yuanhan Zhang, BoLi, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, and Dahua Lin.Mmbench: Is your multi-modal model an all-around player?ArXiv, abs/2307.06281, 2023b.URL https://api.semanticscholar.org/CorpusID:259837088.
  • Liu etal. [2023c]Yuliang Liu, Zhang Li, Hongliang Li, Wenwen Yu, Mingxin Huang, Dezhi Peng, Mingyu Liu, Mingrui Chen, Chunyuan Li, Lianwen Jin, etal.On the hidden mystery of ocr in large multimodal models.arXiv preprint arXiv:2305.07895, 2023c.
  • Mao etal. [2016]Junhua Mao, Jonathan Huang, Alexander Toshev, Oana Camburu, Alan Yuille, and Kevin Murphy.Generation and comprehension of unambiguous object descriptions.In CVPR, 2016.
  • Miller [1995]GeorgeA Miller.Wordnet: a lexical database for english.Communications of the ACM, 38(11):39–41, 1995.
  • Minderer etal. [2024]Matthias Minderer, Alexey Gritsenko, and Neil Houlsby.Scaling open-vocabulary object detection.Advances in Neural Information Processing Systems, 36, 2024.
  • OpenAI [2023]OpenAI.Gpt-4v(ision) system card, 2023.URL https://cdn.openai.com/papers/GPTV_System_Card.pdf.
  • Ordonez etal. [2011]Vicente Ordonez, Girish Kulkarni, and Tamara Berg.Im2text: Describing images using 1 million captioned photographs.Advances in neural information processing systems, 24, 2011.
  • Papineni etal. [2002]Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu.Bleu: a method for automatic evaluation of machine translation.In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318, 2002.
  • Radford etal. [2021]Alec Radford, JongWook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, etal.Learning transferable visual models from natural language supervision.In International conference on machine learning, pages 8748–8763. PMLR, 2021.
  • Rasheed etal. [2023]Hanoona Rasheed, Muhammad Maaz, Sahal Shaji, Abdelrahman Shaker, Salman Khan, Hisham Cholakkal, RaoM Anwer, Erix Xing, Ming-Hsuan Yang, and FahadS Khan.Glamm: Pixel grounding large multimodal model.arXiv preprint arXiv:2311.03356, 2023.
  • Reid etal. [2024]Machel Reid, Nikolay Savinov, Denis Teplyashin, Dmitry Lepikhin, Timothy Lillicrap, Jean-baptiste Alayrac, Radu Soricut, Angeliki Lazaridou, Orhan Firat, Julian Schrittwieser, etal.Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context.arXiv preprint arXiv:2403.05530, 2024.
  • Reimers and Gurevych [2019]Nils Reimers and Iryna Gurevych.Sentence-BERT: Sentence embeddings using Siamese BERT-networks.In Kentaro Inui, Jing Jiang, Vincent Ng, and Xiaojun Wan, editors, Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), November 2019.
  • Sarto etal. [2023]Sara Sarto, Manuele Barraco, Marcella Cornia, Lorenzo Baraldi, and Rita Cucchiara.Positive-augmented contrastive learning for image and video captioning evaluation.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6914–6924, 2023.
  • Schuhmann etal. [2021]Christoph Schuhmann, Richard Vencu, Romain Beaumont, Robert Kaczmarczyk, Clayton Mullis, Aarush Katta, Theo Coombes, Jenia Jitsev, and Aran Komatsuzaki.Laion-400m: Open dataset of clip-filtered 400 million image-text pairs.arXiv preprint arXiv:2111.02114, 2021.
  • Sharma etal. [2018]Piyush Sharma, Nan Ding, Sebastian Goodman, and Radu Soricut.Conceptual captions: A cleaned, hypernymed, image alt-text dataset for automatic image captioning.In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2556–2565, 2018.
  • Tito etal. [2021]Rubèn Tito, Dimosthenis Karatzas, and Ernest Valveny.Document collection visual question answering.In Document Analysis and Recognition–ICDAR 2021: 16th International Conference, Lausanne, Switzerland, September 5–10, 2021, Proceedings, Part II 16, pages 778–792. Springer, 2021.
  • Touvron etal. [2023]Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, etal.Llama 2: Open foundation and fine-tuned chat models.arXiv preprint arXiv:2307.09288, 2023.
  • Vedantam etal. [2015]Ramakrishna Vedantam, CLawrenceZitnick, and Devi Parikh.Cider: Consensus-based image description evaluation.In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4566–4575, 2015.
  • Wang etal. [2023a]Weihan Wang, Qingsong Lv, Wenmeng Yu, Wenyi Hong, JiQi, Yan Wang, Junhui Ji, Zhuoyi Yang, Lei Zhao, Xixuan Song, Jiazheng Xu, Bin Xu, Juanzi Li, Yuxiao Dong, Ming Ding, and Jie Tang.Cogvlm: Visual expert for pretrained language models.ArXiv, abs/2311.03079, 2023a.URL https://api.semanticscholar.org/CorpusID:265034288.
  • Wang etal. [2023b]Weiyun Wang, Min Shi, Qingyun Li, Wenhai Wang, Zhenhang Huang, Linjie Xing, Zhe Chen, Hao Li, Xizhou Zhu, Zhiguo Cao, etal.The all-seeing project: Towards panoptic visual recognition and understanding of the open world.In The Twelfth International Conference on Learning Representations, 2023b.
  • Wang etal. [2024]Weiyun Wang, Yiming Ren, Haowen Luo, Tiantong Li, Chenxiang Yan, Zhe Chen, Wenhai Wang, Qingyun Li, Lewei Lu, Xizhou Zhu, etal.The all-seeing project v2: Towards general relation comprehension of the open world.arXiv preprint arXiv:2402.19474, 2024.
  • Young etal. [2014a]Peter Young, Alice Lai, Micah Hodosh, and Julia Hockenmaier.From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions.Transactions of the Association for Computational Linguistics, 2:67–78, 2014a.doi: 10.1162/tacl_a_00166.URL https://aclanthology.org/Q14-1006.
  • Young etal. [2014b]Peter Young, Alice Lai, Micah Hodosh, and Julia Hockenmaier.From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions.Transactions of the Association for Computational Linguistics, 2:67–78, 2014b.
  • Yu etal. [2023]Weihao Yu, Zhengyuan Yang, Linjie Li, Jianfeng Wang, Kevin Lin, Zicheng Liu, Xinchao Wang, and Lijuan Wang.Mm-vet: Evaluating large multimodal models for integrated capabilities.ArXiv, abs/2308.02490, 2023.URL https://api.semanticscholar.org/CorpusID:260611572.
  • Yue etal. [2024]Xiang Yue, Yuansheng Ni, Kai Zhang, Tianyu Zheng, Ruoqi Liu, GeZhang, Samuel Stevens, Dongfu Jiang, Weiming Ren, Yuxuan Sun, Cong Wei, Botao Yu, Ruibin Yuan, Renliang Sun, Ming Yin, Boyuan Zheng, Zhenzhu Yang, Yibo Liu, Wenhao Huang, Huan Sun, YuSu, and Wenhu Chen.Mmmu: A massive multi-discipline multimodal understanding and reasoning benchmark for expert agi.In Proceedings of CVPR, 2024.

Appendix A Prompt Templates for Detail Caption Benchmark Curation

Prompt of GPT4-Evaluation scores generation.

In order to verify the effect of the proposed CAPTURE metric on a larger evaluation set, we use GPT4[1] instead of humans for evaluation. To better align with human preferences, we manually construct three in context learning cases as shown in Figure 3. In each case, a standard caption and three candidate captions are given, and the corresponding human evaluation results are listed as references, including the relative ranking and the absolute scores. Finally, the current ground truth and candidate captions to be evaluated are given in the same format, prompting GPT4 to output the corresponding evaluation results. And we select the output captions of LLaVA-1.5[35], CogVLM[54] and ShareCaptioner[9] as three candidates for evaluation.

Benchmarking and Improving Detail Image Caption (3)
Prompt of detail caption generation.

In the process of generating detail captions, we use multiple different prompts for GPT-4V[41] to obtain diverse captions as shown in Figure 4. For Gemini-Pro-1.5[46], we found that the model is more likely to output short captions when the prompt does not indicate the expected output length. Based on this, we only use a single prompt with a word limit for generation.

Benchmarking and Improving Detail Image Caption (4)

Appendix B Implementation Details for CAPTURE Metric

Core information extraction.

Core information extraction module aims to extract objects, attributes and relations from a given caption for following matching modules.We adopt a SOTA text scene graph parser: Factual parser[29] as the backbone model.Factual parser is a T5-base model trained on human-annotated scene graph parsing dataset.It takes as input a short caption paragraph, and produce the objects, attributes and relations appearing in the caption.Since Factual parser is trained on short caption parsing dataset, its performance deteriorates severely when given detail image captions.To solve this problem, we first use NLTK toolkit[5] to cut detail image caption into short paragraphs, and apply Factual parser to each paragraph to obtain a list of parsing results.The parsing results are then merged into a larger scene graph based on the following rules:(1) all nouns and adjectives are lemmatized with Wordnet[39];(2) duplicated objects are merged as one element, so are corresponding attributes;(3) attributes describing two or more merged objects are deduplicated;(4) duplicated relations are merged as one element;In this way, we obtain a large scene graph for each caption with duplicated elements removed.The scene graph is then used to compute the final matching score.

Stop words filtering.

Although yielding relatively satisfying parsing results, Factual parser struggles to discriminate concrete nouns from abstract ones, which are not expected to participate in the following matching process.For example, in caption "Two white sheep are enjoying the moment", "sheep" refers to a perceptible element in the image, while "moment" has no tangible meaning.We filter out abstract nouns via a stop word list: once an object in parsing results appear to be in the stop word list, the word itself will not participate in the object elements matching process.

To construct such the stop word list, we first apply LLaMA2-13b-chat[52] and Factual parser to ShareGPT4V-102k dataset for nouns extraction, respectively.We observe that LLaMA may omit a proportion of objects appearing in the caption, but the extracted concrete nouns demonstrate impressive precision.Based on this obeservation, we curate words recalled by Factual parser but omitted by LLaMA, and compute the frequency of these words.Human experts are tasked to judge whether words with the highest frequency are concrete nous or abstract ones.Finally, 500 abstract nouns with the highest frequency are curated to be the stop word list.

It is also worth noticing that although yielding relatively satisfying parsing results, Factual parser struggles when dealing with cross-sentence pronoun reference.When given ambiguous pronoun references, Factual parser may generate objects which are not contained in the caption.To tackle this problem, we further check the parsed objects’ appearance in the caption, and filter out unmatched objects as well as its corresponding attributes and relations.

Core information matching.

After obtaining and filtering core information from both ground truth detail caption and candidate one, the extracted elements are matched to produce final evaluation result.Intuitively, identical object, attribute or relation elements are matched.However, due to the diverse writing style of LVLMs, the same element can be expressed in various ways, and exact matching strategy fail to handle such cases.To solve this problem, we add a synonym matching module after exact matching to match elements with similar meanings.We employ Wordnet to get the synonym set of both the candidate element and ground truth one, and match them if their synonym sets overlaps.Matched candidate objects, attributes and relations are formulated as:

candtypematch=candtypeexcandtypesyn,𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑚𝑎𝑡𝑐𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑒𝑥𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑠𝑦𝑛\begin{split}&cand_{type}^{match}=cand_{type}^{ex}\bigcup cand_{type}^{syn},\\\end{split}start_ROW start_CELL end_CELL start_CELL italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_m italic_a italic_t italic_c italic_h end_POSTSUPERSCRIPT = italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_e italic_x end_POSTSUPERSCRIPT ⋃ italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_s italic_y italic_n end_POSTSUPERSCRIPT , end_CELL end_ROW(1)

where type{obj,attr,rel}𝑡𝑦𝑝𝑒𝑜𝑏𝑗𝑎𝑡𝑡𝑟𝑟𝑒𝑙{type}\in\{obj,attr,rel\}italic_t italic_y italic_p italic_e ∈ { italic_o italic_b italic_j , italic_a italic_t italic_t italic_r , italic_r italic_e italic_l }. candtypeex𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑒𝑥cand_{type}^{ex}italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_e italic_x end_POSTSUPERSCRIPT and candtypesyn𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑠𝑦𝑛cand_{type}^{syn}italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_s italic_y italic_n end_POSTSUPERSCRIPT stand for exactly matched and synonym matched candidate phrases, respectively.Matched ground truth elements are formulated in the same way as gtobjmatch𝑔superscriptsubscript𝑡𝑜𝑏𝑗𝑚𝑎𝑡𝑐gt_{obj}^{match}italic_g italic_t start_POSTSUBSCRIPT italic_o italic_b italic_j end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_m italic_a italic_t italic_c italic_h end_POSTSUPERSCRIPT, gtattrmatch𝑔superscriptsubscript𝑡𝑎𝑡𝑡𝑟𝑚𝑎𝑡𝑐gt_{attr}^{match}italic_g italic_t start_POSTSUBSCRIPT italic_a italic_t italic_t italic_r end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_m italic_a italic_t italic_c italic_h end_POSTSUPERSCRIPT and gtrelmatch𝑔superscriptsubscript𝑡𝑟𝑒𝑙𝑚𝑎𝑡𝑐gt_{rel}^{match}italic_g italic_t start_POSTSUBSCRIPT italic_r italic_e italic_l end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_m italic_a italic_t italic_c italic_h end_POSTSUPERSCRIPT.

Exact matching and synonym matching strategies tackle most of the matched cases, but still fail to cover all core information extracted from captions in various writing styles.To this end, we propose a soft matching strategy, which takes Sentence BERT[47] model to encode remaining object, attribute or relation phrases and compute a matching score in [0,1)01[0,1)[ 0 , 1 ) for remaining unmatched phrases.Let candtyperm𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑟𝑚cand_{type}^{rm}italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT be unmatched candidate phrases and gttyperm𝑔superscriptsubscript𝑡𝑡𝑦𝑝𝑒𝑟𝑚gt_{type}^{rm}italic_g italic_t start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT be ground truth ones, their similarity matrix Styperm𝐑|candtyperm|×|gttyperm|superscriptsubscript𝑆𝑡𝑦𝑝𝑒𝑟𝑚superscript𝐑𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑟𝑚𝑔superscriptsubscript𝑡𝑡𝑦𝑝𝑒𝑟𝑚S_{type}^{rm}\in\mathbf{R}^{|cand_{type}^{rm}|\times|gt_{type}^{rm}|}italic_S start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT ∈ bold_R start_POSTSUPERSCRIPT | italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT | × | italic_g italic_t start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT | end_POSTSUPERSCRIPT is calculated as:

Styperm=ϕ(candtyperm)×ϕ(gttyperm)T,superscriptsubscript𝑆𝑡𝑦𝑝𝑒𝑟𝑚italic-ϕ𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑟𝑚italic-ϕsuperscript𝑔superscriptsubscript𝑡𝑡𝑦𝑝𝑒𝑟𝑚𝑇S_{type}^{rm}=\phi(cand_{type}^{rm})\times\phi(gt_{type}^{rm})^{T},italic_S start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT = italic_ϕ ( italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT ) × italic_ϕ ( italic_g italic_t start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT ) start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT ,(2)

where ϕ()italic-ϕ\phi(\cdot)italic_ϕ ( ⋅ ) denotes Sentence BERT model.We further compute the matching score of candtyperm𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑟𝑚cand_{type}^{rm}italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT and gttyperm𝑔superscriptsubscript𝑡𝑡𝑦𝑝𝑒𝑟𝑚gt_{type}^{rm}italic_g italic_t start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT as follows:

cand_matchtyperm[i]=maxj=1,2,,|gttyperm|Styperm[i,j],gt_matchtyperm[j]=maxi=1,2,,|candtyperm|Styperm[i,j].formulae-sequence𝑐𝑎𝑛𝑑_𝑚𝑎𝑡𝑐superscriptsubscript𝑡𝑦𝑝𝑒𝑟𝑚delimited-[]𝑖subscript𝑗12𝑔superscriptsubscript𝑡𝑡𝑦𝑝𝑒𝑟𝑚superscriptsubscript𝑆𝑡𝑦𝑝𝑒𝑟𝑚𝑖𝑗𝑔𝑡_𝑚𝑎𝑡𝑐superscriptsubscript𝑡𝑦𝑝𝑒𝑟𝑚delimited-[]𝑗subscript𝑖12𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑟𝑚superscriptsubscript𝑆𝑡𝑦𝑝𝑒𝑟𝑚𝑖𝑗\begin{split}&cand\_match_{type}^{rm}[i]=\max_{j=1,2,...,|gt_{type}^{rm}|}S_{%type}^{rm}[i,j],\\&gt\_match_{type}^{rm}[j]=\max_{i=1,2,...,|cand_{type}^{rm}|}S_{type}^{rm}[i,j%].\\\end{split}start_ROW start_CELL end_CELL start_CELL italic_c italic_a italic_n italic_d _ italic_m italic_a italic_t italic_c italic_h start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT [ italic_i ] = roman_max start_POSTSUBSCRIPT italic_j = 1 , 2 , … , | italic_g italic_t start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT | end_POSTSUBSCRIPT italic_S start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT [ italic_i , italic_j ] , end_CELL end_ROW start_ROW start_CELL end_CELL start_CELL italic_g italic_t _ italic_m italic_a italic_t italic_c italic_h start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT [ italic_j ] = roman_max start_POSTSUBSCRIPT italic_i = 1 , 2 , … , | italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT | end_POSTSUBSCRIPT italic_S start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT [ italic_i , italic_j ] . end_CELL end_ROW(3)

cand_matchtyperm𝑐𝑎𝑛𝑑_𝑚𝑎𝑡𝑐superscriptsubscript𝑡𝑦𝑝𝑒𝑟𝑚cand\_match_{type}^{rm}italic_c italic_a italic_n italic_d _ italic_m italic_a italic_t italic_c italic_h start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT and gt_matchtyperm𝑔𝑡_𝑚𝑎𝑡𝑐superscriptsubscript𝑡𝑦𝑝𝑒𝑟𝑚gt\_match_{type}^{rm}italic_g italic_t _ italic_m italic_a italic_t italic_c italic_h start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT are then used as a complementary to exact matched and synonym matched relations.

After obtaining matching results, we compute the precision and recall for each type of core information.The precision and recall are computed as:

precisiontype=|candtypematch||candtype|,recalltype=|gttypematch||gttype|.formulae-sequence𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜subscript𝑛𝑡𝑦𝑝𝑒𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑚𝑎𝑡𝑐𝑐𝑎𝑛subscript𝑑𝑡𝑦𝑝𝑒𝑟𝑒𝑐𝑎𝑙subscript𝑙𝑡𝑦𝑝𝑒𝑔superscriptsubscript𝑡𝑡𝑦𝑝𝑒𝑚𝑎𝑡𝑐𝑔subscript𝑡𝑡𝑦𝑝𝑒\begin{split}&precision_{type}=\frac{|cand_{type}^{match}|}{|cand_{type}|},\\&recall_{type}=\frac{|gt_{type}^{match}|}{|gt_{type}|}.\\\end{split}start_ROW start_CELL end_CELL start_CELL italic_p italic_r italic_e italic_c italic_i italic_s italic_i italic_o italic_n start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT = divide start_ARG | italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_m italic_a italic_t italic_c italic_h end_POSTSUPERSCRIPT | end_ARG start_ARG | italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT | end_ARG , end_CELL end_ROW start_ROW start_CELL end_CELL start_CELL italic_r italic_e italic_c italic_a italic_l italic_l start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT = divide start_ARG | italic_g italic_t start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_m italic_a italic_t italic_c italic_h end_POSTSUPERSCRIPT | end_ARG start_ARG | italic_g italic_t start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT | end_ARG . end_CELL end_ROW(4)

Attribute precision and recall are computed in the same way.As for relation elements, candidate matching score and ground truth matching score are counted separately due to the introduction of feature matching:

precisiontype=|candtypematch|+cand_matchtyperm|cand_matchtyperm||candtype|,recalltype=|gttypematch|+gt_matchtyperm|gt_matchtyperm||gttype|.formulae-sequence𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜subscript𝑛𝑡𝑦𝑝𝑒𝑐𝑎𝑛superscriptsubscript𝑑𝑡𝑦𝑝𝑒𝑚𝑎𝑡𝑐𝑐𝑎𝑛𝑑_𝑚𝑎𝑡𝑐superscriptsubscript𝑡𝑦𝑝𝑒𝑟𝑚𝑐𝑎𝑛𝑑_𝑚𝑎𝑡𝑐superscriptsubscript𝑡𝑦𝑝𝑒𝑟𝑚𝑐𝑎𝑛subscript𝑑𝑡𝑦𝑝𝑒𝑟𝑒𝑐𝑎𝑙subscript𝑙𝑡𝑦𝑝𝑒𝑔superscriptsubscript𝑡𝑡𝑦𝑝𝑒𝑚𝑎𝑡𝑐𝑔𝑡_𝑚𝑎𝑡𝑐superscriptsubscript𝑡𝑦𝑝𝑒𝑟𝑚𝑔𝑡_𝑚𝑎𝑡𝑐superscriptsubscript𝑡𝑦𝑝𝑒𝑟𝑚𝑔subscript𝑡𝑡𝑦𝑝𝑒\begin{split}&precision_{type}=\frac{|cand_{type}^{match}|+\frac{\sum{cand\_%match_{type}^{rm}}}{|cand\_match_{type}^{rm}|}}{|cand_{type}|},\\&recall_{type}=\frac{|gt_{type}^{match}|+\frac{\sum{gt\_match_{type}^{rm}}}{|%gt\_match_{type}^{rm}|}}{|gt_{type}|}.\\\end{split}start_ROW start_CELL end_CELL start_CELL italic_p italic_r italic_e italic_c italic_i italic_s italic_i italic_o italic_n start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT = divide start_ARG | italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_m italic_a italic_t italic_c italic_h end_POSTSUPERSCRIPT | + divide start_ARG ∑ italic_c italic_a italic_n italic_d _ italic_m italic_a italic_t italic_c italic_h start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT end_ARG start_ARG | italic_c italic_a italic_n italic_d _ italic_m italic_a italic_t italic_c italic_h start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT | end_ARG end_ARG start_ARG | italic_c italic_a italic_n italic_d start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT | end_ARG , end_CELL end_ROW start_ROW start_CELL end_CELL start_CELL italic_r italic_e italic_c italic_a italic_l italic_l start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT = divide start_ARG | italic_g italic_t start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_m italic_a italic_t italic_c italic_h end_POSTSUPERSCRIPT | + divide start_ARG ∑ italic_g italic_t _ italic_m italic_a italic_t italic_c italic_h start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT end_ARG start_ARG | italic_g italic_t _ italic_m italic_a italic_t italic_c italic_h start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_m end_POSTSUPERSCRIPT | end_ARG end_ARG start_ARG | italic_g italic_t start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT | end_ARG . end_CELL end_ROW(5)

Finally, CAPTURE metric takes the precision and recall of all three types of core information into consideration, and produce the final evaluation result as:

CAPTURE=αF1obj+βF1attr+γF1relα+β+γ,𝐶𝐴𝑃𝑇𝑈𝑅𝐸𝛼𝐹subscript1𝑜𝑏𝑗𝛽𝐹subscript1𝑎𝑡𝑡𝑟𝛾𝐹subscript1𝑟𝑒𝑙𝛼𝛽𝛾CAPTURE=\frac{\alpha F1_{obj}+\beta F1_{attr}+\gamma F1_{rel}}{\alpha+\beta+%\gamma},italic_C italic_A italic_P italic_T italic_U italic_R italic_E = divide start_ARG italic_α italic_F 1 start_POSTSUBSCRIPT italic_o italic_b italic_j end_POSTSUBSCRIPT + italic_β italic_F 1 start_POSTSUBSCRIPT italic_a italic_t italic_t italic_r end_POSTSUBSCRIPT + italic_γ italic_F 1 start_POSTSUBSCRIPT italic_r italic_e italic_l end_POSTSUBSCRIPT end_ARG start_ARG italic_α + italic_β + italic_γ end_ARG ,(6)

where α𝛼\alphaitalic_α, β𝛽\betaitalic_β and γ𝛾\gammaitalic_γ are scale factors, and F1type=precisiontyperecalltypeprecisiontype+recalltype𝐹subscript1𝑡𝑦𝑝𝑒𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜subscript𝑛𝑡𝑦𝑝𝑒𝑟𝑒𝑐𝑎𝑙subscript𝑙𝑡𝑦𝑝𝑒𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜subscript𝑛𝑡𝑦𝑝𝑒𝑟𝑒𝑐𝑎𝑙subscript𝑙𝑡𝑦𝑝𝑒F1_{type}=\frac{precision_{type}\cdot recall_{type}}{precision_{type}+recall_{%type}}italic_F 1 start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT = divide start_ARG italic_p italic_r italic_e italic_c italic_i italic_s italic_i italic_o italic_n start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT ⋅ italic_r italic_e italic_c italic_a italic_l italic_l start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT end_ARG start_ARG italic_p italic_r italic_e italic_c italic_i italic_s italic_i italic_o italic_n start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT + italic_r italic_e italic_c italic_a italic_l italic_l start_POSTSUBSCRIPT italic_t italic_y italic_p italic_e end_POSTSUBSCRIPT end_ARG stands for the F1 score of each type of core information.

Appendix C Visualized Examples for Improved Detail Caption Construction

Cases of detail caption construction.
Benchmarking and Improving Detail Image Caption (5)

In Figure 5, we show the effectiveness of detail caption construction in Section 4.1 with three visualized cases.In the first case, highlighted in red, the LVLM-generated caption incorrectly mentions that there are people in the image, while the caption produced by our pipeline removes the relevant description correctly.In the following two cases, the synthesized captions complement model-generated captions with additional visual information highlighted in green, resulting into higher-quality detail image caption.

Benchmarking and Improving Detail Image Caption (2024)

References

Top Articles
Latest Posts
Article information

Author: Jonah Leffler

Last Updated:

Views: 5994

Rating: 4.4 / 5 (65 voted)

Reviews: 88% of readers found this page helpful

Author information

Name: Jonah Leffler

Birthday: 1997-10-27

Address: 8987 Kieth Ports, Luettgenland, CT 54657-9808

Phone: +2611128251586

Job: Mining Supervisor

Hobby: Worldbuilding, Electronics, Amateur radio, Skiing, Cycling, Jogging, Taxidermy

Introduction: My name is Jonah Leffler, I am a determined, faithful, outstanding, inexpensive, cheerful, determined, smiling person who loves writing and wants to share my knowledge and understanding with you.