  <?xml version="1.0"?>
<journal>
 <journal_metadata>
  <full_title>Journal of Intelligent Systems and Internet of Things</full_title>
  <abbrev_title>JISIoT</abbrev_title>
  <issn media_type="print">2690-6791</issn>
  <issn media_type="electronic">2769-786X</issn>
  <doi_data>
   <doi>10.54216/JISIoT</doi>
   <resource>https://www.americaspg.com/journals/show/4175</resource>
  </doi_data>
 </journal_metadata>
 <journal_issue>
  <publication_date media_type="print">
   <year>2019</year>
  </publication_date>
  <publication_date media_type="online">
   <year>2019</year>
  </publication_date>
 </journal_issue>
 <journal_article publication_type="full_text">
  <titles>
   <title>Real-Time Gesture Recognition Using Attention-Based CNN-RNN Framework for Human-Robot Interaction</title>
  </titles>
  <contributors>
   <organization sequence="first" contributor_role="author">Assistant Professor, School of Computer Science Engineering, SRM Institute of Science and Technology, Ramapuram, Chennai, Tamilnadu, India</organization>
   <person_name sequence="first" contributor_role="author">
    <given_name>R.</given_name>
    <surname>R.</surname>
   </person_name>
   <organization sequence="first" contributor_role="author">Assistant Professor, Department of Computer Science and Engineering, Vel Tech Rangarajan Dr. Sagunthala R&amp;D Institute of Science and Technology, Avadi, Chennai, Tamil Nadu, India</organization>
   <person_name sequence="additional" contributor_role="author">
    <given_name>Chinnathambi</given_name>
    <surname>Kamatchi</surname>
   </person_name>
   <organization sequence="first" contributor_role="author">Assistant Professor, Department of Electronics and Instrumentation Engineering, Sri Ramakrishna Engineering College, Coimbatore, Tamil Nadu, India</organization>
   <person_name sequence="additional" contributor_role="author">
    <given_name>Y.</given_name>
    <surname>Dharshan</surname>
   </person_name>
   <organization sequence="first" contributor_role="author">Assistant Professor, Department of Electronics and Communication Engineering, Hindusthan Institute of Technology, Coimbatore, Tamil Nadu, India</organization>
   <person_name sequence="additional" contributor_role="author">
    <given_name>K.</given_name>
    <surname>Kowsalya</surname>
   </person_name>
   <organization sequence="first" contributor_role="author">Assistant Professor, Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation (Deemed to be University), Andhra Pradesh, India</organization>
   <person_name sequence="additional" contributor_role="author">
    <given_name>R.</given_name>
    <surname>Vijay</surname>
   </person_name>
   <organization sequence="first" contributor_role="author">Professor, Department of Artificial Intelligence and Data Science, Dr. Mahalingam College of Engineering and Technology, Pollachi, Coimbatore, Tamil Nadu, India</organization>
   <person_name sequence="additional" contributor_role="author">
    <given_name>M.</given_name>
    <surname>Balakrishnan</surname>
   </person_name>
  </contributors>
  <jats:abstract xml:lang="en">
   <jats:p>Gesture recognition serves as a key enabler for natural and intuitive human–robot interaction (HRI) in smart automation and assistive systems. However, achieving real-time performance with high recognition accuracy remains a significant challenge due to dynamic background variations, occlusion, and complex spatio-temporal dependencies in gesture sequences. This paper presents a real-time attention-based CNN-RNN framework for robust gesture recognition and adaptive HRI in dynamic environments. The proposed system utilizes Convolutional Neural Networks (CNNs) for spatial feature extraction from sequential video frames and Bidirectional Recurrent Neural Networks (BiRNNs)—integrated with an attention mechanism—for modeling temporal dependencies and focusing on discriminative motion cues. The attention layer enhances interpretability by prioritizing salient gestures and reducing background noise. A hybrid optimization strategy, combining adaptive learning rate scheduling and regularized dropout, ensures computational stability and generalization across gesture datasets. Experiments conducted on benchmark datasets such as NVIDIA Dynamic Gesture (NvGesture) and ChaLearn IsoGD demonstrate superior performance, achieving an accuracy of 97.8% and a real-time inference speed of 34 FPS, outperforming baseline CNN, 3D-CNN, and LSTM architectures. The proposed framework effectively balances accuracy, latency, and interpretability, making it suitable for real-world HRI applications, including service robotics, industrial automation, and assistive technologies.</jats:p>
  </jats:abstract>
  <publication_date media_type="print">
   <year>2025</year>
  </publication_date>
  <publication_date media_type="online">
   <year>2025</year>
  </publication_date>
  <pages>
   <first_page>398</first_page>
   <last_page>408</last_page>
  </pages>
  <doi_data>
   <doi>10.54216/JISIoT.170128</doi>
   <resource>https://www.americaspg.com/articleinfo/18/show/4175</resource>
  </doi_data>
 </journal_article>
</journal>
