Aug. 11, 2022, 1:12 a.m. | Youyuan Zhang, Jiuniu Wang, Hao Wu, Wenjia Xu

cs.CV updates on arXiv.org arxiv.org

Image captioning models are usually trained according to human annotated
ground-truth captions, which could generate accurate but generic captions. In
this paper, we focus on generating the distinctive captions that can
distinguish the target image from other similar images. To evaluate the
distinctiveness of captions, we introduce a series of metrics that use
large-scale vision-language pre-training model CLIP to quantify the
distinctiveness. To further improve the distinctiveness of captioning models,
we propose a simple and effective training strategy which trains …

arxiv captioning clip cv image optimization

Data Scientist (m/f/x/d)

@ Symanto Research GmbH & Co. KG | Spain, Germany

Associate Data Engineer

@ Redkite | London, England, United Kingdom

Data Management Associate Consultant

@ SAP | Porto Salvo, PT, 2740-262

NLP & Data Modelling Consultant - SAP LABS

@ SAP | Bengaluru, IN, 560066

Catalog Data Quality Specialist

@ Delivery Hero | Montevideo, Uruguay

Data Analyst for CEO Office with Pathway to Functional Analyst

@ Amar Bank | Jakarta