  <?xml version="1.0"?>
<journal>
 <journal_metadata>
  <full_title>Journal of Intelligent Systems and Internet of Things</full_title>
  <abbrev_title>JISIoT</abbrev_title>
  <issn media_type="print">2690-6791</issn>
  <issn media_type="electronic">2769-786X</issn>
  <doi_data>
   <doi>10.54216/JISIoT</doi>
   <resource>https://www.americaspg.com/journals/show/4054</resource>
  </doi_data>
 </journal_metadata>
 <journal_issue>
  <publication_date media_type="print">
   <year>2019</year>
  </publication_date>
  <publication_date media_type="online">
   <year>2019</year>
  </publication_date>
 </journal_issue>
 <journal_article publication_type="full_text">
  <titles>
   <title>Advanced Deep Learning Model for Image Captioning Using Customized Vision Transformer with Global Optimization Algorithm</title>
  </titles>
  <contributors>
   <organization sequence="first" contributor_role="author">Department of Computer Engineering, College of Computer Engineering &amp; Sciences, Prince Sattam bin Abdulaziz University, Alkharj-11942, Saudi Arabia</organization>
   <person_name sequence="first" contributor_role="author">
    <given_name>Suleman</given_name>
    <surname>Suleman</surname>
   </person_name>
   <organization sequence="first" contributor_role="author">Department of Computer Engineering, College of Computer Engineering &amp; Sciences, Prince Sattam bin Abdulaziz University, Alkharj-11942, Saudi Arabia</organization>
   <person_name sequence="additional" contributor_role="author">
    <given_name>Mohammed Altaf</given_name>
    <surname>Ahmed</surname>
   </person_name>
  </contributors>
  <jats:abstract xml:lang="en">
   <jats:p>In the image-captioning field, the excellence of produced captions is vital for the effectual interaction of visual content. Image Captioning is the main task, which unites computer vision (CV) and natural language processing (NLP), where it goals to produce graphic legends for images. A dual-fold procedure depends on precise image perception and alters language understanding both semantically and syntactically. It is gradually challenging to stay up with the modern study and consequences in image captioning owing to the developing amount of knowledge accessible on the topic.  This analysis examines into deep learning (DL) to tackle the tasks challenged by individuals with graphic impairments, targeting to improve their visual insight via advanced technologies. By tradition, the visually impaired have trusted physical support and adaptive helps for understanding and navigating visual content. With the beginning of DL, there is a unique chance to develop this scenery. In this paper, we offer an Advanced Deep Learning Method for Image Captioning Based Using Customized Transformer with a Global Optimization Algorithm (ADLIC-CTGOA). The foremost aim of ADLIC-CTGOA model is to focus on the initiation of the effectual textual image captioning of an input image. Initially, the ADLIC-CTGOA method employs preprocessing phase to enhances both image and text data: images undergo noise removal and contrast enhancement to improve quality, while text is processed by removing numbers, converting to lowercase, and text vectorization. Next, the customized swin transformer is employed for feature extraction to capture fine-grained visual features from images. In addition, the BERT Transformer model is deployed for image captioning process. To enhance the performance of proposed technique, the chaotic Aquila optimization (CAO) technique was applied for parameter tuning for enhancing the performance. A wide sort of simulation studies are executed to ensure the improved performance of ADLIC-CTGOA system. The comparative result exploration reported the betterment of the ADLIC-CTGOA model on recent approaches in terms of different evaluation measures.</jats:p>
  </jats:abstract>
  <publication_date media_type="print">
   <year>2026</year>
  </publication_date>
  <publication_date media_type="online">
   <year>2026</year>
  </publication_date>
  <pages>
   <first_page>273</first_page>
   <last_page>289</last_page>
  </pages>
  <doi_data>
   <doi>10.54216/JISIoT.180219</doi>
   <resource>https://www.americaspg.com/articleinfo/18/show/4054</resource>
  </doi_data>
 </journal_article>
</journal>
