Image Caption Combined with GAN Training Method - Archive ouverte HAL Access content directly
Conference Papers Year : 2020

Image Caption Combined with GAN Training Method

(1, 2) , (2)
1
2

Abstract

In today’s world where the number of images is huge and people cannot quickly retrieve the information they need, we urgently need a simpler and more human-friendly way of understanding images, and image captions have emerged. Image caption, as its name suggests, is to analyze and understand image information to generate natural language descriptions of specific images. In recent years, it has been widely used in image-text crossover studies, early infant education, and assisted by disadvantaged groups. And the favor of industry, has produced many excellent research results. At present, the evaluation of image caption is basically based on objective evaluation indicators such as BLUE and CIDEr. It is easy to prevent the generated caption from approaching human language expression. The introduction of GAN idea allows us to use a new method of adversarial training. To evaluate the generated caption, the evaluation module is more natural and comprehensive. Considering the requirements for image fidelity, this topic proposes a GAN-based image description. The Attention mechanism is introduced to improve image fidelity, which makes the generated caption more accurate and more close to human language expression.
Fichier principal
Vignette du fichier
498234_1_En_29_Chapter.pdf (377.32 Ko) Télécharger le fichier
Origin : Files produced by the author(s)

Dates and versions

hal-03456989 , version 1 (30-11-2021)

Licence

Attribution - CC BY 4.0

Identifiers

Cite

Zeqin Huang, Zhongzhi Shi. Image Caption Combined with GAN Training Method. 11th International Conference on Intelligent Information Processing (IIP), Jul 2020, Hangzhou, China. pp.310-316, ⟨10.1007/978-3-030-46931-3_29⟩. ⟨hal-03456989⟩
14 View
0 Download

Altmetric

Share

Gmail Facebook Twitter LinkedIn More