Published June 1, 2025 | Version v1
Journal article Open

Transformer+transformer architecture for image captioning in Indonesian language

  • 1. Calvin Institute of Technology

Description

Image captioning in Indonesian language poses a significant challenge due to the complex interplay between visual and linguistic comprehension, as well as the scarcity of publicly available datasets. Despite considerable advancements in this field, research specifically targeting the Indonesian language remains scarce. In this paper, we propose a novel image captioning model employing a transformer-based architecture for both the encoder and decoder components. Our model is trained and evaluated on the pre-translated Flickr30k dataset in the Indonesian language. We conduct acomparative analysis of various transformer configurations and convolutional neural network (CNN)-recurrent neural network (RNN) architectures. Our findings highlight the superior performance of a vision transformer (ViT) as the visual encoder, combined with IndoBERT as the textual decoder. This architecture achieved a BLEU-4 score of 0.223 and a ROUGE-L score of 0.472.

Files

62 26891.pdf

Files (3.8 MB)

Name Size Download all
md5:dbd9a5c2869d5dd883f760639af4a7c1
3.8 MB Preview Download