Classifier model for lecturer evaluation by students using speech emotion recognition and deep learning approaches
Abstract
Lecturers play a crucial role in higher education, with their teaching behavior directly impacting learning and teaching quality. Lecturer evaluation by students (LES) is a common method for assessing lecturer performance, though it often relies on subjective perceptions. As a more objective alternative, speech emotion recognition (SER) uses speech technology to analyze emotions in the speech of lecturers during classes. This study proposes using deep learning-based SER, including convolutional neural network (CNN) and bidirectional long short-term memory (Bi-LSTM), to evaluate teaching quality by analyzing displayed emotions. Removing silence from audio signals is crucial for enhancing feature analysis, such as energy, zero-crossing rate (ZCR), and mel-frequency cepstral coefficients (MFCC). This method removes inactive segments, emphasizing significant segments, and improving accuracy in detecting voice and emotions. Results show that the 1D CNN model with Bi-LSTM, using MFCC with 13 coefficients, energy, and ZCR, performs excellently in emotion detection, achieving a validation accuracy of over 0.851 with an accuracy gap of 0.002. This small gap indicates good generalization and reduces the risk of overfitting, making teaching evaluations more objective and valuable for improving practices.
Keywords
Bi-LSTM; Energy; Evaluation; Lecturer; MFCC; Student; Zero-crossing rate
Full Text:
PDFDOI: http://doi.org/10.11591/ijai.v14.i6.pp5157-5171
Refbacks
- There are currently no refbacks.
Copyright (c) 2025 Yesy Diah Rosita, Wahyu Andi Saputra

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
IAES International Journal of Artificial Intelligence (IJ-AI)
ISSN/e-ISSN 2089-4872/2252-8938
This journal is published by the Institute of Advanced Engineering and Science (IAES).