BRIDGE: Bridging Gaps in Image Captioning Evaluation with Stronger Visual Cues

July 29, 2024 ยท Entered Twilight ยท ๐Ÿ› European Conference on Computer Vision

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: README.md, images

Authors Sara Sarto, Marcella Cornia, Lorenzo Baraldi, Rita Cucchiara arXiv ID 2407.20341 Category cs.CV: Computer Vision Cross-listed cs.AI, cs.CL, cs.MM Citations 12 Venue European Conference on Computer Vision Repository https://github.com/aimagelab/bridge-score โญ 13 Last Checked 1 month ago
Abstract
Effectively aligning with human judgment when evaluating machine-generated image captions represents a complex yet intriguing challenge. Existing evaluation metrics like CIDEr or CLIP-Score fall short in this regard as they do not take into account the corresponding image or lack the capability of encoding fine-grained details and penalizing hallucinations. To overcome these issues, in this paper, we propose BRIDGE, a new learnable and reference-free image captioning metric that employs a novel module to map visual features into dense vectors and integrates them into multi-modal pseudo-captions which are built during the evaluation process. This approach results in a multimodal metric that properly incorporates information from the input image without relying on reference captions, bridging the gap between human judgment and machine-generated image captions. Experiments spanning several datasets demonstrate that our proposal achieves state-of-the-art results compared to existing reference-free evaluation scores. Our source code and trained models are publicly available at: https://github.com/aimagelab/bridge-score.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision