Home > Published Issues > 2024 > Volume 15, No. 4, 2024 >
JAIT 2024 Vol.15(4): 532-543
doi: 10.12720/jait.15.4.532-543

Automatic Gender Authentication from Arabic Speech Using Hybrid Learning

Amjad Rehman Khan
Artificial Intelligence and Data Analytics Lab (AIDA), College of Computer and Information Sciences,
Prince Sultan University, Riyadh 11586, Saudi Arabia
E-mail: arkhan@psu.edu.sa

Manuscript received April 1, 2023; revised July 16, 2023, accepted August 14, 2023; published April 24, 2024.

Abstract—Speech recognition is progressively being utilized in practical applications with time. Automatic gender identification is one of the most intriguing applications since it distinguishes female and male speeches from briefly spoken communication records. This is advantageous in various applications, including automated conversation systems, system verification, demographic attribute prediction and assessing speaker’s expressions. Speech is a natural mode of communication, and pitch variation of a gender-specific speech signal is often used to identify a person as male or female. This paper presents a model to identify gender from Arabic speech by integrating audio preprocessing, Mel-Frequency Cepstral Coefficients (MFCC), Delta MFCC, and Log Filter bank feature extraction. Pre-processing involves testing pre-emphasis, framing, windowing, and Fast Fourier Transform. Finally, features are extracted using three feature extraction methods from the processed audios. Feed Forward Neural Networks and Keras-based Neural Networks are employed as classifier models. Regarding accuracy and simplicity, the proposed hybrid method surpasses most previous approaches discussed in the literature for gender categorization from Arabic speech. The proposed model achieved an average classification accuracy of 93.09%.
 
Keywords—speech recognition, Arabic language, gender classification, hybrid learning, technological development

Cite: Amjad Rehman Khan, "Automatic Gender Authentication from Arabic Speech Using Hybrid Learning," Journal of Advances in Information Technology, Vol. 15, No. 4, pp. 532-543, 2024.

Copyright © 2024 by the authors. This is an open access article distributed under the Creative Commons Attribution License (CC BY-NC-ND 4.0), which permits use, distribution and reproduction in any medium, provided that the article is properly cited, the use is non-commercial and no modifications or adaptations are made.