[1]
T. Jaiswal, M. Pandey, and P. Tripathi, "Image captioning through cognitive iot and machine-learning approaches", Turkish J. Comput. Math. Edu., vol. 12, no. 9, pp. 333-351, 2021.
[7]
A. Radford, L. Metz, and S. Chintala, "Unsupervised representation learning with deep convolutional generative adversarial networks", In 4th Int. Conf. Learn. Represent. ICLR 2016 - Conf. Track Proc, 2015. https://api.semanticscholar.org/CorpusID:11758569
[10]
Q. Zheng, P. Zhao, Y. Li, H. Wang, and Y. Yang, "Spectrum interference-based two-level data augmentation method in deep learning for automatic modulation classification., vol. Vol. 33", In Neural Comput. Appl., 2020, pp. 7723-7745
[11]
Q. Zheng, M. Yang, Q. Zhang, and X. Zhang, "Fine-grained image classification based on the combination of artificial features and deep convolutional activation features", In 2017 IEEE/CIC Int. Conf. Commun, China, 2017, pp. 1-6
[19]
Y. Yang, C.L. Teo, H. Daumé, and Y. Aloimonos, "Corpus-guided sentence generation of natural images", In EMNLP 2011 - Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference, year, 2011, pp. 444-454
[23]
J. Mao, W. Xu, Y. Yang, J. Wang, and A.L. Yuille, "Explain images with multimodal recurrent neural networks", In ArXiv, vol. 1410, 2014, p. 1
[24]
K. Xu, "Show, attend and tell: Neural image caption generation with visual attention", In 32nd Int. Conf. Mach. Learn. ICML 2015, vol. 3, 2015, pp. 2048-2057
[26]
I. Schwartz, A.G. Schwing, and T. Hazan, "High-order attention models for visual question answering", Adv. Neural Inf. Process. Syst., pp. 3665-3675, 2017.
[29]
M. Mirza, and S. Osindero, "Conditional Generative Adversarial Nets", ArXiv1411.1, 2014.
[32]
L. Meng, J. Wang, Y. Yang, and L. Xiao, "Prior Knowledge-Guided Transformer for Remote Sensing Image Captioning", In IEEE Trans., vol. Vol. 61. Geosci. Remote Sens, 2023, pp. 1-13
[33]
T. Chen, Z. Li, J. Wu, H. Ma, and B. Su, "Improving image captioning with Pyramid Attention and SC-GAN., vol. Vol. 117", In Image Vis. Comput., 2021, p. 104340
[34]
P. Song, D. Guo, J. Zhou, M. Xu, and M. Wang, "Memorial GAN With Joint Semantic Optimization for Unpaired Image Captioning", In IEEE Trans., vol. Vol. 53. Cybern, 2022, pp. 4388-4399
[36]
X. Yang, H. Zhang, C. Gao, and J. Cai, "Learning to Collocate Visual-Linguistic Neural Modules for Image Captioning., vol. Vol. 131", In Int. J. Comput. Vis., 2022, pp. 82-100
[38]
K. Simonyan, and A. Zisserman, "Very deep convolutional networks for large-scale image recognition", In 3rd Int. Conf. Learn. Represent. ICLR 2015 - Conf. Track Proc, 2014.
[42]
K. Papineni, S. Roukos, T. Ward, and W.J. Zhu, "BLEU: A method for automatic evaluation of machine translation", In Proceedings of the Annual Meeting of the Association for Computational Linguistics year, 2002, pp. 311-318
[43]
C.Y. Lin, "Rouge: A package for automatic evaluation of summaries", In Proceedings of the workshop on text summarization branches out (WAS 2004) year., 2004, pp. 25-26
[44]
S. Banerjee, and A. Lavie, "METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgmentsin", In IEEvaluation@ ACL., 2005, pp. 65-72
[54]
M. Z. Hossain, F. Sohel, M. F. Shiratuddin, and H. Shiratuddin, "Text to image synthesis for improved image captioning", In IEEE Access year., vol. 9, 2021, pp. 64918-64928
[55]
Y. Zhou, W. Tao, and W. Zhang, "Triple Sequence Generative Adversarial Nets for Unsupervised Image Captioning", In ICASSP 2021 - 2021 IEEE Int.. Conf. Acoust. Speech Signal Process, 2021, pp. 7598-7602