Sentiment Analysis of Student Feedback Using Attention-Based RNN and Transformer Embedding
Sentiment Analysis of Student Feedback Using Attention-Based RNN and Transformer Embedding
Corresponding Author:
Imad Zyout
Department of Computer and Communication Engineering, College of Engineering
Tafila Technical University
Tafila, Jordan
Email: [email protected]
1. INTRODUCTION
In the realm of natural language processing (NLP), sentiment analysis (SA) serves as a crucial tool
developed to extract the emotional content hidden within user feedback [1], [2]. SA algorithms have been
developed to handle text data collected from various sources, including social media, educational online sites,
healthcare records, and student reviews [3]. Feedback represents the response of end-users, whether online
or offline, regarding provided services or their level of satisfaction [4]. SA techniques play a vital role in the
development and enhancement of both commercial [5] and educational services [6]. User feedback, whether
provided online or offline, reflects their satisfaction with the services received, particularly in the realm of
education. SA models have been employed to examine course evaluations and student feedback regarding
their satisfaction with courses and teachers. Similarly, SA can be employed in other settings where users offer
feedback on electronic services [5]. Analyzing reviews by SA is important for service providers to identify
areas for improvement and enhance the overall user experience, ultimately resulting in higher user satisfaction
[7], [8]. The application of SA in the educational sector is mainly to enhance the learning experience for both
teachers and students [9]. With the advent of advanced educational services and distance learning platforms,
the integration of SA into the field of education has become increasingly relevant [10]. SA systems are built
to collect and analyze students’ responses to evaluate their satisfaction with various aspects of the education
subject, including the teacher, assignments, and exams [11].
The process of SA and the extraction of feelings, from the text, is done in the form of different levels.
These levels include entity level, where sentiment or opinion analysis is conducted on feedback related to an
educational entity [12]; sentence level, where the sentiment of the document towards a particular context is
analyzed to ascertain if it is positive or negative [13]; document level, where it is determined if the document
expresses a positive or negative sentiment towards a given context [14]; and aspect level, providing insight
into the positive or negative aspects of educational practices [15]. Particularly, in the application of SA to the
education sector, educators can gain valuable insights into student sentiment, which can inform and improve
the design and delivery of educational content and services [16].
The key stages to create an effective educational SA model are,as the first stage, the collection and
labeling of a feedback dataset [17]. The second stage involves various prepossessing steps, such as cleaning,
tokenizing, removing stop words, and stemming. Additionally, the words are encoded using modeling tech-
niques such as term frequency-inverse document frequency (TF-IDF) and term frequency (TF), with the n-gram
should be selected before embedding these words into machine learning (ML) algorithms [18], [19]. SA tech-
niques rely on the use of classical supervised ML and modern algorithms to categorize student feedback into
positive, negative, or neutral sentiments. Traditional ML algorithms including support vector machines (SVM),
decision trees (DT), random forest (RF), and naive Bayes (NB) are often used alone or combined to develop
the SA model [3], [19]. In addition, ML techniques such as voting, ensemble, and bagging are employed to im-
prove the accuracy of SA models in the field of education. Furthermore, advanced ML including deep learning
techniques like long short-term memory (LSTM) networks [7] and convolutional neural networks (CNN) are
utilized to create both supervised and unsupervised SA models for educational purposes [20]. SA approaches
that utilize deep learning techniques employ embedding methods to normalize a sequence of vectors into a
fixed dimension. The word2vec embedding technique, which utilizes neural networks (NN) to vectorize words
into a single vector based on a large corpus of words in a context, has been widely used in SA [21]. Another
popular embedding technique is global vectors (GloVe) for word representation, which depends on the co-
occurrence of words in a given context [22]. In recent years, embedded models such as bidirectional encoder
representations from transformers (BERT)-base [23], robustly optimized bidirectional encoder representations
from transformers approach (RoBERTa)-base [24], and a lite bidirectional encoder representations from trans-
formers (ALBERT)-base [25] have been developed based on self-attention mechanisms to retain the position
of words in the context. These models obtain vectors by surrounding the words of the objective word and have
shown remarkable performance in various NLP tasks, including sentiment analysis [26], [27].
In higher education, the sentiment analysis of student comments, leverages advances in artificial intel-
ligence (AI) and NLP fields to adequately extract constructive feedback and opinions; and learning aspects. A
representative SA model can improve and tune teaching methodology to be more effective and to suit different
groups of learners. Efforts to build SA models have used different methods to pre-process text and applied var-
ious NLP tools including text-cleaning, -tokenizing, and text vectorization methods. Research studies also vary
in modeling the opinion-mining problem and in the approach utilized to accomplish the opinion-recognition
task. Following the application of different text processing and vectorization methods, the classification prob-
lem was solved using divers techniques including classical and modern ML algorithms. Pallathadka et al. [28]
suggested to forecast the student’s performance using ML algorithms including NB, ID3, C4.5, and SVM.
On an online dataset, from UCI, to test these four models, the SVM achieved the highest accuracy of 89%.
Toçoğlu and Onan et al. [29] conducted a sentiment analysis study on Turkish student reviews using various
ML algorithms (SVM, NB, logistic regression (LR), RF, AdaBoost, bagging, and the voting algorithm) and text
vectorization methods. The study showed that the models based on TF-IDF outperformed the other approaches
with scores ranging from 57% to 73%. Okoye et al. [30] developed an educational process and data mining
plus machine learning (EPDM + ML) model that uses text mining and k-nearest neighbor (KNN) algorithm
to analyze teacher performance based on student evaluations. Their analysis showed that 76.4% of the student
comments they analyzed were predominantly positive, while 23.6% contained some kind of positive or neg-
ative sentiment. The study also found that female students are more likely to recommend teachers based on
sentiment, with a precision, recall, specificity, accuracy, and F1-score of 100%, while males are slightly more
influenced by emotion with a precision of 94.4%, recall of 100%, specificity of 97.3%, accuracy of 97.3%, and
F1-score of 97.1%. The EPDM+ML model was shown to be an effective predictor of student recommendations
for teachers with a zero error rate, indicating its potential usefulness in educational settings. Faizi and El Fkihi
[31] applied SA to classify positive and negative student reviews collected from course evaluations and social
media platforms. Using the SVM algorithm, they achieved an accuracy score of 93.35%. Lalata et al. [32]
used SA to analyze student comments in the classroom. The ensemble and individual models of LR, SVM, DT,
and RF algorithms were evaluated and compared on a dataset of comments of 1413 positive, 327 negative, and
82 neutral comments. The algorithm-based voting method produced the best accuracy of 90.32%. Rakhmanov
[33] compared different text vectorization models such as TF-IDF and counter vector for analyzing students’
comments. They built a sentiment model and compared several ML algorithms including RF, SVM, NB, gra-
dient boosting, and artificial neural network (ANN). The models were evaluated using 55,000 TF-IDF features
extracted from student comments. The results showed that the RF-based TF-IDF method was the most effec-
tive, achieving an accuracy of 97%. Sindhu et al. [34] utilized multiple feature extraction including TF-IDF,
true false (TF), and true positive (TP) with n-grams ranging from 1 to 3 and several ML techniques to analyze
massive open online courses (MOOCs) reviews.
Several studies have recently utilized modern ML methods, particularly, deep learning techniques.
The first application of deep learning for evaluating faculty teaching performance from students’ feedback was
presented in [35]. The study presented the supervised aspect-based opinion mining system based on a two-
layered LSTM model. On two datasets including a manually tagged dataset and a standard SemEval-2014
dataset and utilizing the domain embedding, the proposed system achieves high accuracy rates of 91% and
93% for both aspect extraction and sentiment polarity detection tasks, respectively. Onan [35] combined CNN,
recurrent neural network (RNN), LSTM, and gated recurrent unit (GRU) and three embedding techniques:
word2vec, GloVe, and FastText, were using 66,000 MOOC online reviews and the LSTM-based GloVe embed-
ding achieved an accuracy of 95.8%. Yousafzai et al. [36] integrated the bidirectional long short-term memory
(Bi-LSTM) NN with an attention mechanism and a feature selection method to forecast student performance.
The proposed approach attained a 91% accuracy rate on the UCI dataset, utilizing 33 features to characterize a
student’s behavior throughout a course. A new stemming algorithm was presented in [37] to enhance the SA
accuracy of Hausa language, a widely spoken in West Africa. The SA of Hausa language was done classical
and modern ML techniques including transformer approaches such as BERT and RoBERTa. Authors devel-
oped a monolingual large corpus dataset of about 40,000 Hausa-English comments, named the HESAC. Using
the new stemming and achieving about 97.4% by applying the suggested algorithm during the pre-processing
phase, which was slightly better than the cross-lingual approach.
This paper presents the SA approach, which is intended to determine the sentiment of student com-
ments about their courses and teachers. The researchers manually collected and labeled comments expressing
opinions about courses. To address the limitations of a small dataset of negative comments, the large language
model (LLM) and generative AI, specifically ChatGPT, were used to synthesize additional comments. This
project utilizes prompt engineering, the design of specific prompts to guide the output of LLM, and generates a
wide spectrum of student feedback, including exam question difficulty, teaching style, and fairness of grading.
To the best of the author’s knowledge, this is the first work, in the context of SA of customer review, that lever-
ages generative AI to build the development dataset. This work also explored several SA model architectures
based on deep learning algorithms such as RNN, GRU, LSTM, and Bi-LSTM. To optimize the performance of
the proposed SA model, Glove and Bert were utilized to improve the model’s performance on the test dataset.
The best network was selected through multiple experiments and its performance was optimized using attention
layers.
2. METHODS
The methodology for developing the SA approach can be described in three stages. Firstly, the dataset
is preprocessed to clean and transform the data. Secondly, the SA models are developed and trained on the
preprocessed data. Finally, the performance of the models is evaluated and compared using various metrics.
Figure 1 illustrates the three stages of the SA model development process.
Sentiment analysis of student feedback using attention-based RNN and transformer embedding (Imad Zyout)
2176 ❒ ISSN: 2252-8938
Sentiment analysis of student feedback using attention-based RNN and transformer embedding (Imad Zyout)
2178 ❒ ISSN: 2252-8938
The architecture of the SA-classical and SA-GloVe models utilizes a variety of multi-NN, including
simple RNN, GRU, LSTM, and Bi-LSTM. Two embedding approaches are adopted: one with an embedded
vector space having random weights (max length 500), and the other using initial weights from the GloVe
dictionaries. Each model consists of a single layer with 100 neurons and integrates both dropout and spatial
dropout-1D to minimize overfitting. Post feature extraction in a 100-dimensional space, the softmax activation
function determines the probability ratio between positive and negative classes.
The SA-RNN model had an embedding layer with parameters (word length, 300, embedding matrix,
500), SpatialDropout1D of 0.3, RNN layer with 100 neurons, and a Dense layer with 2 neurons. The SA-GRU
model had an embedding layer with parameters (word length, 300, embedding matrix, 500), SpatialDropout1D
of 0.3, a GRU layer with 100 neurons, and a dense layer with 2 neurons. The SA−LSTM model had an
embedding layer with parameters (word length, 300, embedding matrix, 500), Dropout of 0.25, LSTM layer
with 100 neurons, and a dense layer with 2 neurons. Finally, the SA-Bi−LSTM model had an embedding layer
with parameters (word length, 300, embedding matrix, 500), dropout and recurrent dropout of 0.3 and 0.25,
respectively, a Bi−LSTM layer with 100 neurons, and a dense layer with 2 neurons.
Table 2 displays the second set of SA models, utilizing Keras’ BERT implementation. These models
employ NN (RNN, GRU, LSTM, and Bi-LSTM) similar to classical SA models but differ in using BERT layers
for word embeddings. BERT, pre-trained on extensive text data, encodes complex linguistic relationships. Each
BERT-based SA model includes one layer with dropout and SpatialDropout1D to prevent overfitting. Extracted
features (100 dimensions) from reviews are input to prediction layers using softmax activation. Overall, BERT-
based SA models exhibit superior performance compared to classical models, highlighting the advantages of
utilizing pre-trained language models like BERT for text classification tasks.
-1 and 1 using the Tanh activation function and then computed using the softmax activation function to create a
set of weights. These weights are used to create context vectors by concatenating the input data with the output
weights and multiplying them with the input features. The resulting vectors are then added together to create
the final form of the context vector. Both SA-Bi-LSTM and SA-Bi-LSTM-BERT models use this attention
layer to optimize the prediction scores for the two approaches. The output of the attention layer is input into a
dense layer to find the prediction score for each class.
Sentiment analysis of student feedback using attention-based RNN and transformer embedding (Imad Zyout)
2180 ❒ ISSN: 2252-8938
Overall, the results suggest that using GloVe embeddings can improve the performance of SA models.
However, it’s important to note that the differences in performance between the models were relatively small
and other factors such as the size and composition of the dataset, or the specific parameters used in each model,
could also have an impact on performance.
The high performance of these models is primarily attributed to the use of BERT embedding. BERT
models were pre-trained on a large corpus of text data and have shown to be highly effective in NLP tasks. Our
results suggest that the SA model, utilizing BERT embedding, excels in accurately identifying sentiment in stu-
dents’ reviews. This achievement has important implications for educational institutes, particularly regarding
course reviews and instructors’ training.
The SA model, lacking an embedding method, exhibited significant inaccuracies, misclassifying in-
stances like ”maybe need to chill a bit” as positive, contrary to annotators and BERT-based SA models. The
Sentiment analysis of student feedback using attention-based RNN and transformer embedding (Imad Zyout)
2182 ❒ ISSN: 2252-8938
phrase ”found it hard to complete the assignment on time” was inaccurately labeled by classical networks and
GloVe embedding but correctly identified as negative by BERT-based SA. Optimization attempts for GloVe
within SA models, including LSTM, GRU, Bi-LSTM, and RNN, aimed to enhance performance but fell short
of overcoming inherent limitations, resulting in non-negligible error rates. Challenges in accurately classifying
phrases, like ”the teacher seems easily frustrated,” persisted. Notable improvement was seen in the attention
layer of BERT-based SA, as illustrated in Figure 5. SA models (Bi-LSTM-attention) and (Bi-LSTM-BERT-
attention) outperformed, achieving exceptional predictions with 155 and 160 instances correctly classified in
the negative class, and 525 and 532 instances in the positive class, respectively.
4. CONCLUSION
The use of SA has become increasingly popular in analyzing people’s opinions across various domain
mains. In the educational context, SA can be used to gauge student satisfaction and demand for teachers’
services. However, such models face limitations due to small database sizes and the diverse deep learning
tools used. To overcome this, this proposed a SA system using three methods, including automatic embedding
and GloVe and BERT embedding, to analyze a database of student chat categorized into positive and negative
feedback. Classical networks such as RNN, GRU, LSTM, and Bi-LSTM were employed, with and without
pre-trained GloVe embeddings, achieving F-scores ranging from 67% to 69%. A BERT Keras embedding
layer-based sentiment model was also evaluated, achieving F-scores ranging from 83% to 87%. The addition
of an attention layer to the Bi-LSTM SA model yielded the highest performance, resulting in an enhanced
F-score of 89% for the Bi-LSTM-BERT sentiment model and 88% for the Bi-LSTM sentiment model. The
proposed system can thus provide valuable insights into student satisfaction and demand for teachers’ services,
contributing to the enhancement of educational quality. Though the results of this work revealed that the
SA model architecture, indeed, has influenced the overall performance, the data-centric approach necessitates
further research into its potential to boost the model performance.
REFERENCES
[1] T. Shaik, X. Tao, C. Dann, H. Xie, Y. Li, and L. Galligan, “Sentiment analysis and opinion mining on educational data: A survey,”
Natural Language Processing Journal, vol. 2, pp. 1–11, Mar. 2023, doi: 10.1016/j.nlp.2022.100003.
[2] M. Alzyout, E. AL Bashabsheh, H. Najadat, and A. Alaiad, “Sentiment Analysis of Arabic Tweets about Violence Against Women
using Machine Learning,” in 2021 12th International Conference on Information and Communication Systems (ICICS), 2021, pp.
171–176, doi: 10.1109/ICICS52457.2021.9464600.
[3] D. D. Dsouza, Deepika, D. P. Nayak, E. J. Machado, and N. D. Adesh, “Sentimental analysis of student feedback using machine
learning techniques,” International Journal of Recent Technology and Engineering, vol. 8, no. 1 Special Issue 4, pp. 986–991, 2019.
[4] S. Ulfa, R. Bringula, C. Kurniawan, and M. Fadhli, “Student Feedback on Online Learning by Using Sentiment Analy-
sis: A Literature Review,” in 2020 6th International Conference on Education and Technology (ICET), 2020, pp. 53–58, doi:
10.1109/ICET51153.2020.9276578.
[5] R. Baragash and H. Aldowah, “Sentiment analysis in higher education: A systematic mapping review,” Journal of Physics: Confer-
ence Series, vol. 1860, no. 1, pp. 1–13, 2021, doi: 10.1088/1742-6596/1860/1/012002.
[6] D. K. Dake and E. Gyimah, “Using sentiment analysis to evaluate qualitative students’ responses,” Education and Information
Technologies, vol. 28, no. 4, pp. 4629–4647, 2023, doi: 10.1007/s10639-022-11349-1.
[7] I. A. Kandhro, S. Wasi, K. Kumar, M. Rind, and M. Ameen, “Sentiment Analysis of Students Comment by using Long-Short Term
Model,” Indian Journal of Science and Technology, vol. 12, no. 8, pp. 1–16, 2019, doi: 10.17485/ijst/2019/v12i8/141741.
[8] L. K. Singh, “An Review of Student Sentimental Analysis For Educational Database Using Unsupervised Machine Learning Ap-
proaches.,” European Journal of Molecular & Clinical Medicine, vol. 7, no. 9, pp. 2151–2165, 2021.
[9] A. Tzacheva and A. Easwaran, “Emotion Detection and Opinion Mining from Student Comments for Teaching Innovation Assess-
ment,” International Journal of Education (IJE), vol. 9, no. 2, pp. 21–32, 2021, doi: 10.5121/ije2021.9203.
[10] J. Zhou and J. Ye, “Sentiment analysis in education research: a review of journal publications,” Interactive Learning Environments,
vol. 31, no. 3, pp. 1252–1264, 2023, doi: 10.1080/10494820.2020.1826985.
[11] F. Dalipi, K. Zdravkova, and F. Ahlgren, “Sentiment Analysis of Students’ Feedback in MOOCs: A Systematic Literature Review,”
Frontiers in Artificial Intelligence, vol. 4, pp. 1–13, 2021, doi: 10.3389/frai.2021.728708.
[12] R. Yang, “Machine Learning and Deep Learning for Sentiment Analysis over Students’ Reviews: An Overview Study,” Preprints,
pp. 1-9, 2021.
[13] M. Sivakumar and U. S. Reddy, “Aspect based sentiment analysis of students opinion using machine learning techniques,” in 2017
International Conference on Inventive Computing and Informatics (ICICI), 2017, pp. 726–731, doi: 10.1109/ICICI.2017.8365231.
[14] F. S. Dolianiti et al., “Sentiment analysis on educational datasets: a comparative evaluation of commercial tools,” Educational
Journal of the University of Patras UNESCO Chair, vol. 6, no. 1, pp. 262–273, 2019.
[15] T. Shaik et al., “A Review of the Trends and Challenges in Adopting Natural Language Processing Methods for Education Feedback
Analysis,” IEEE Access, vol. 10, pp. 56720–56739, 2022, doi: 10.1109/ACCESS.2022.3177752.
[16] D. Wehbe, A. Alhammadi, H. Almaskari, K. Alsereidi, and H. Ismail, “UAE e-Learning Sentiment Analysis Framework,” in The
7th Annual International Conference on Arab Women in Computing in Conjunction with the 2nd Forum of Women in Research,
2021, pp. 1–4, doi: 10.1145/3485557.3485570.
[17] K. Nilanga, M. Herath, H. Maduwantha, and S. Ranathunga, “Dataset and Baseline for Automatic Student Feedback Analysis,” in
Proceedings of the 13th Conference on Language Resources and Evaluation (LREC 2022), 2022, pp. 2042–2049.
[18] Z. Nasim, Q. Rajput, and S. Haider, “Sentiment analysis of student feedback using machine learning and lexicon based ap-
proaches,” in 2017 International Conference on Research and Innovation in Information Systems (ICRIIS), 2017, pp. 1–6, doi:
10.1109/ICRIIS.2017.8002475.
[19] C. A. Pacol and T. D. Palaoag, “Enhancing Sentiment Analysis of Textual Feedback in the Student-Faculty Evaluation using
Machine Learning Techniques,” European Journal of Engineering Science and Technology, vol. 4, no. 1, pp. 27–34, 2021, doi:
10.33422/ejest.v4i1.604.
[20] I. A. Kandhro, M. A. Chhajro, K. Kumar, H. N. Lashari, and U. Khan, “Student Feedback Sentiment Analysis Model Using
Various Machine Learning Schemes A Review,” Indian Journal of Science and Technology, vol. 14, no. 12, pp. 1–9, 2019, doi:
10.17485/ijst/2019/v12i14/143243.
[21] T. Mikolov, K. Chen, G. Corrado, and J. Dean, “Efficient estimation of word representations in vector space,” in 1st International
Conference on Learning Representations, ICLR 2013 - Workshop Track Proceedings, 2013, pp. 1–12.
[22] J. Pennington, R. Socher, and C. Manning, “Glove: Global Vectors for Word Representation,” in Proceedings of the 2014 Conference
on Empirical Methods in Natural Language Processing (EMNLP), 2014, pp. 1532–1543, doi: 10.3115/v1/D14-1162.
[23] J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understand-
ing,” in NAACL HLT 2019 - 2019 Conference of the North American Chapter of the Association for Computational Linguistics:
Human Language Technologies, 2019, vol. 1, pp. 4171–4186.
[24] Y. Liu et al., “RoBERTa: A Robustly Optimized BERT Pretraining Approach,” in ICLR 2020 Conference, 2019, pp. 1–15.
[25] Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma, and R. Soricut, “Albert: A lite bert for self-supervised learning of language
representations,” in International Conference on Learning Representations, 2020, pp. 1–17.
[26] R. Bensoltane and T. Zaki, “Towards Arabic aspect-based sentiment analysis: a transfer learning-based approach,” Social Network
Analysis and Mining, vol. 12, no. 1, 2022, doi: 10.1007/s13278-021-00794-4.
[27] R. Mao, Q. Liu, K. He, W. Li, and E. Cambria, “The Biases of Pre-Trained Language Models: An Empirical Study on Prompt-Based
Sentiment Analysis and Emotion Detection,” IEEE Transactions on Affective Computing, vol. 14, no. 3, pp. 1743–1753, 2023, doi:
10.1109/TAFFC.2022.3204972.
[28] H. Pallathadka, A. Wenda, E. Ramirez-Ası́s, M. Ası́s-López, J. Flores-Albornoz, and K. Phasinam, “Classification and prediction of
student performance data using various machine learning algorithms,” Materials Today: Proceedings, vol. 80, pp. 3782–3785, 2023,
doi: 10.1016/j.matpr.2021.07.382.
[29] M. A. Toçoğlu and A. Onan, “Sentiment Analysis on Students’ Evaluation of Higher Educational Institutions,” in Intelligent and
Fuzzy Techniques: Smart and Innovative Solutions, Cham: Springer, 2021, pp. 1693–1700, doi: 10.1007/978-3-030-51156-2 197.
[30] K. Okoye, A. Arrona-Palacios, C. Camacho-Zuñiga, J. A. G. Achem, J. Escamilla, and S. Hosseini, “Towards teaching analytics: a
contextual model for analysis of students’ evaluation of teaching through text mining and machine learning classification,” Education
and Information Technologies, vol. 27, no. 3, pp. 3891–3933, 2022, doi: 10.1007/s10639-021-10751-5.
[31] R. Faizi and S. El Fkihi, “A Sentiment Analysis Based Approach for Exploring Student Feedback,” in Innovative Technologies and
Learning, Cham: Springer, 2022, pp. 52–59, doi: 10.1007/978-3-031-15273-3 6.
[32] J. A. P. Lalata, B. Gerardo, and R. Medina, “A sentiment analysis model for faculty comment evaluation using ensemble machine
learning algorithms,” in ACM International Conference Proceeding Series, 2019, pp. 68–73, doi: 10.1145/3341620.3341638.
[33] O. Rakhmanov, “A Comparative Study on Vectorization and Classification Techniques in Sentiment Analysis to Classify Student-
Lecturer Comments,” Procedia Computer Science, vol. 178, pp. 194–204, 2020, doi: 10.1016/j.procs.2020.11.021.
[34] I. Sindhu, S. Muhammad Daudpota, K. Badar, M. Bakhtyar, J. Baber, and M. Nurunnabi, “Aspect-Based Opinion Mining on Stu-
dent’s Feedback for Faculty Teaching Performance Evaluation,” IEEE Access, vol. 7, pp. 108729–108741, 2019, doi: 10.1109/AC-
CESS.2019.2928872.
Sentiment analysis of student feedback using attention-based RNN and transformer embedding (Imad Zyout)
2184 ❒ ISSN: 2252-8938
[35] A. Onan, “Sentiment analysis on massive open online course evaluations: A text mining and deep learning approach,” Computer
Applications in Engineering Education, vol. 29, no. 3, pp. 572–589, 2021, doi: 10.1002/cae.22253.
[36] B. K. Yousafzai et al., “Student-performulator: Student academic performance using hybrid deep neural network,” Sustainability,
vol. 13, no. 17, pp. 1–21, 2021, doi: 10.3390/su13179775.
[37] O. Rakhmanov and T. Schlippe, “Sentiment Analysis for Hausa: Classifying Students’ Comments,” in 1st Annual Meeting of the
ELRA/ISCA Special Interest Group on Under-Resourced Languages, SIGUL 2022 - held in conjunction with the International
Conference on Language Resources and Evaluation, LREC 2022, 2022, pp. 98–105.
BIOGRAPHIES OF AUTHORS