Skip to main content

A hybrid approach for binary and multi-class classification of voice disorders using a pre-trained model and ensemble classifiers

Abstract

Recent advances in artificial intelligence-based audio and speech processing have increasingly focused on the binary and multi-class classification of voice disorders. Despite progress, achieving high accuracy in multi-class classification remains challenging. This paper proposes a novel hybrid approach using a two-stage framework to enhance voice disorders classification performance, and achieve state-of-the-art accuracies in multi-class classification. Our hybrid approach, combines deep learning features with various powerful classifiers. In the first stage, high-level feature embeddings are extracted from voice data spectrograms using a pre-trained VGGish model. In the second stage, these embeddings are used as input to four different classifiers: Support Vector Machine (SVM), Logistic Regression (LR), Multi-Layer Perceptron (MLP), and an Ensemble Classifier (EC). Experiments are conducted on a subset of the Saarbruecken Voice Database (SVD) for male, female, and combined speakers. For binary classification, VGGish-SVM achieved the highest accuracy for male speakers (82.45% for healthy vs. disordered; 75.45% for hyperfunctional dysphonia vs. vocal fold paresis), while VGGish-EC performed best for female speakers (71.54% for healthy vs. disordered; 68.42% for hyperfunctional dysphonia vs. vocal fold paresis). In multi-class classification, VGGish-SVM outperformed other models, achieving mean accuracies of 77.81% for male speakers, 63.11% for female speakers, and 70.53% for combined genders. We conducted a comparative analysis against related works, including the Mel frequency cepstral coefficient (MFCC), MFCC-glottal features, and features extracted using the wav2vec and HuBERT models with SVM classifier. Results demonstrate that our hybrid approach consistently outperforms these models, especially in multi-class classification tasks. The results show the feasibility of a hybrid framework for voice disorder classification, offering a foundation for refining automated tools that could support clinical assessments with further validation.

Peer Review reports

Introduction

Voice production is the process by which humans produce sound to communicate ideas, meaning, opinions, and other information. The human voice production system includes the lungs, larynx, vocal tract, and vocal folds. The lungs provide the air pressure that is needed to vibrate the vocal folds. The vocal folds are located in the larynx, also known as the voice box. When the vocal folds vibrate, they create sound waves that travel through the vocal tract, which is the passage of air from the larynx to the mouth and nose. The shape of the vocal tract affects the timbre of the voice. Voice disorders can occur when there is a problem with any of the components of the voice production system causing changes in the pitch, loudness, or quality of the voice. These disorders can reduce the clarity of a person’s oral communication ability. Voice disorders can vary in severity from minor hoarseness or alterations in vocal quality to the extreme outcome of complete voice loss [1].

Voice disorders can result from various factors. These disorders are commonly classified based on their underlying causes, which may include psychogenic, functional, or organic factors. Organic voice disorders are caused by structural or neurological problems that affect the vocal folds or other parts of the voice production system [2]. Functional voice disorders occur when the vocal mechanism is not used efficiently, even though the physical structure of the larynx and vocal tract is normal. Psychogenic voice disorders, on the other hand, stem from psychological factors such as sadness, anxiety, or emotional responses to traumatic or stressful situations [3].

Voice disorders can have a significant impact on people of all ages, potentially leading to stress, embarrassment, frustration, withdrawal, and depression. Professions that require frequent and demanding use of the voice, such as teaching, acting, and singing, are particularly susceptible to these disorders [4]. To ensure the right treatment, accurate classification of the voice disorders is crucial. A speech therapist typically evaluates the patient’s voice quality for this purpose. However, this approach is subjective and relies on the speech therapist’s expertise. Another approach to assess voice disorders is to use artificial intelligence (AI) to process acoustic features of voice signals, which provides an objective assessment. Automatic classification of voice disorders can provide speech therapists with a faster and more comfortable way to identify voice disorders in patients.

Recent advances in AI have enabled significant progress in audio and speech processing tasks, including speaker identification, speech emotion recognition, and voice disorder detection. For instance, Xie et al. [5] employed attention-based long short-term memory (LSTM) networks to classify speech emotions. Similarly, Keser and Gezer [6] conducted a comparative analysis of speaker identification methods, combining deep learning, machine learning, and subspace classifiers with diverse feature extraction techniques. Authors in [7] further demonstrate the potential of deep multiple instance learning for voice activity detection (VAD). These studies highlight the versatility of hybrid approaches—integrating feature engineering with classifiers like support vector machines (SVMs), logistic regression, and neural networks—to address the unique challenges of audio classification. However, voice disorder classification poses distinct difficulties due to the subtle acoustic variations between disorders and the need for high diagnostic precision. While existing works often focus on binary classification, multi-class frameworks remain underexplored. This gap motivates our proposed framework, which combines deep learning-based feature extraction with robust classifiers to improve both binary and multi-class classification performance.

Related work

In recent years, researchers have become increasingly interested in the automatic classification of voice disorders. These methods, utilizing computer algorithms to analyze speech signals, can revolutionize the classification and detection of voice disorders, making it more objective, efficient, and accessible. Most researchers focus on binary classification problems, such as classifying between healthy and pathological voices or the detection of a voice disorder. Spectrograms and cepstral analysis are two commonly used features for this purpose. However, in recent years, machine learning algorithms have gained popularity for their ability to learn and recognize patterns in acoustic features associated with various types of voice disorders. Fang et al. [8] used Mel Frequency Cepstral Coefficients (MFCCs) and three classifiers, namely Deep Neural Network (DNN), Gaussian Mixture Model (GMM) and SVM, for pathological voice detection. Cordeiro et al. [9] applied hierarchical classification for the identification of pathological voice, employing MFCCs and line spectral frequencies features. Kodrasi et al. [10] proposed a hierarchical multi-class automatic technique using handcrafted acoustic features to distinguish between speech apraxia, dysarthria and neurotypical speech. The approach utilizes two SVMs, with the first SVM distinguishing between neurotypical speech and impaired, while the second SVM discriminates between dysarthria and apraxia of speech.

Costa et al. [11] proposed combining the hidden Markov model (HMM) and modified MFCCs for the voice disorders caused by a vocal fold pathology. In [12, 13], the authors applied multilayer neural networks for the classification of MFCC features and demonstrated that results can be enhanced by considering the differentiation of the speaker’s gender. Ali et al. [14] introduced a method for the classification and detection of voice disorders, utilizing a Gaussian mixture model (GMM) classifier with running speech voice data. Benba et al. [15] investigated the detection of dysphonia using a Naive Bayes (NB) algorithm. They extracted acoustic features using MFCC. Authors in [16] used MFCC features to differentiate between Parkinson’s disease (PD) and healthy voices. They extracted MFCC features from three different vowel sounds: /a/, /o/, and /u/. Authors in [17, 18] also explored binary classification of voice disorders in their research. In [19], the authors employ wavelet scattering features to capture both time-frequency information from voice signals, which are then used for classifying neurological voice disorders.

In addition to binary classification, the majority of research studies use sustained vowel /a/ recordings from clinical settings for their investigations [20]. In [21], the introduction of continuous speech and vowel /a/ analysis for voice disorder identification is discussed. The authors conducted a comparison of glottal features extracted from the sustained vowel sound /a/ and voiced segments within continuous speech. Fujimura et al. [22] used an end-to-end 1D-CNN model to classify voice disorders using voice samples of the sustained vowel /a/. The research demonstrated that the 1D-CNN models were capable of consistently evaluating voice disorders, aligning with human assessments.

In recent years, deep learning has achieved impressive results in a variety of areas, including natural language processing, computer vision and audio analysis. Deep learning’s ability to handle complex and high-dimensional acoustic features makes it well-suited for addressing the challenges of voice disorders classification. This has encouraged many researchers to explore the potential of deep learning for voice disorder classification. Wu et al. [23] developed a novel system using spectrograms of disordered and normal speech recordings as input. They employed Convolutional Deep Belief Networks for pre-training CNN weights as a generative model to understand the input data’s structure statistically. Subsequently, they fine-tuned the CNN using supervised back-propagation. In [24], authors propose the use of a CNN model along with short-time Fourier transform (STFT) features for the binary classification of voice disorders. Mohammed et al. [25] addressed the problem of voice disorder detection by using CNN model. They specifically focused on the automatic detection of depression from speech. Chaiani et al. [26], the authors analyzed an algorithm that extracts a chromagram acoustic feature from voice samples and uses it as input to a CNN-based classification system. The research in [27] proposed a two-stage framework for the classification of different voice disorders. The first stage uses speech enhancement to improve the voice signal quality by removing noise. The second stage employs a CNN with long short-term memory (CNN-LSTM) to learn complex features from spectrograms of the enhanced voice signals. Harar et al. [28] proposed a novel approach for voice pathology detection that uses convolutional and LSTM layers to learn directly from raw audio signals. Furthermore, recent studies have highlighted the beneficial impact of denoising for audio signals [29], advanced vocal feature extraction [30]. These approaches collectively suggest promising avenues for enhancing the automatic classification of voice disorders.

In voice disorders classification, limited data availability is a common challenge. To address this, some researchers have used pre-trained models [31,32,33,34]. In [35], the authors proposed a transfer learning framework that uses a pre-trained OpenL3-SVM model and linear local tangent space alignment (LLTSA) for dimensionality reduction. They first extracted the Mel spectrum of the voice signals and then fed it into the OpenL3 model to obtain high-level feature embeddings. Violeta et al. [36] investigated the performance of self-supervised pre-trained Wav2Vec 2.0 and WavLM models for automatic pathological speech recognition using different setups. Zhu et al. [37, 38] introduced pre-trained BERT and WavBERT models for the detection of dementia using human speech. Karaman et al. [39] employed the SqueezeNet1_1, ResNet101, and DenseNet161 networks for the detection of Parkinson’s disease based on speech signals. The findings showed that the proposed networks, which utilize pre-trained models with a fine-tuning approach, achieved promising results. In [40], the authors used a pre-trained ResNet50 model for dysarthric speech detection.

Research gap and contribution

Most studies on the automatic classification of voice disorders have focused on the binary classification, typically distinguishing between pathological and healthy voices. Some studies have taken a more specialized approach, aiming to identify particular pathological voices among all other pathological and healthy voices. A few studies have investigated multi-class classification of voice disorders, but the accuracy of these approaches is low. Multi-class classification of voice disorders is a challenging problem due to the limited training data and subtle differences between different types of disordered voices. In this study, we address both binary and multi-class classification of voice disorders. For binary classification, we distinguish between healthy and disordered voices, as well as between two different types of pathological voices. For multi-class classification, we have three classes: healthy, vocal fold paresis and hyperfunctional dysphonia.

Gender-specific classification of voice disorders has not been widely investigated. We present classification results separately for male and female speakers, as well as combined results, for both binary and multi-class tasks. This enables us to analyze and compare gender-based differences in the classification of voice disorders.

Feature extraction is a crucial step in machine learning tasks, and it holds particular significance in the classification of voice disorders due to the small dataset size. We utilize the pre-trained VGGish model [41] to extract 128-dimensional high-level embedding features using logarithmic mel spectrogram of voice data. As the name indicates, the VGGish network takes inspiration from a well-known VGG network and is adapted for audio classification. This model was trained on a large Audio set, which was a preliminary version of the YouTube-8M dataset. These embeddings are then utilized as input for machine learning classifiers.

Previous studies have employed transformer-based models like wav2vec and HuBERT for extracting audio embeddings. While these models perform well in general speech tasks, we found that VGGish, a CNN-based model, delivers better results for the classification of voice disorders. Consequently, our approach outperforms transformer-based models in this domain.

Our dataset is imbalanced, which mirrors the distribution often seen in real-world applications, where certain voice disorders are less common. This imbalance presents challenges in accurately classifying minority classes. To overcome this issue, we employed ensemble classifiers that combine the strengths of multiple models, improving performance on minority classes and enhancing overall classification accuracy.

We tested three machine learning classifiers: Logistic Regression (LR), Multi-Layer Perceptron (MLP), and Support Vector Machine (SVM). We also employed an ensemble classifier (EC) using SVM, LR, and MLP with soft voting to combine the predictions of the three classifiers. This allowed us to leverage the collective insights of these diverse classifiers and improve the overall classification performance.

This study demonstrates the effectiveness of utilizing embeddings from a pre-trained VGGish model and ensemble classifiers for both binary and multi-class classification of voice disorders. Additionally, we examine the impact of gender on the classification task. Our findings are compared to popular baseline methods, providing a comprehensive evaluation of our approach. The results show that our method outperforms the baseline approaches on both binary and multi-class classification tasks, demonstrating the superiority of the proposed method.

Paper outline

The rest of the paper is structured as follows. Section “Proposed method” presents an in-depth explanation of the proposed method. Section “Dataset and experimental setup” describes the experimental setup and the voice dataset used in this study. Section “Experiments and results” provides a comprehensive overview of the experiments conducted for both binary and multi-class classification tasks. We present the results and performance metrics achieved by our approach. Section “Discussion” presents the implications of our findings. Section “Conclusion” summarizes the key points and highlights the main contributions made by our study.

Proposed method

In this paper, we propose a novel hybrid two-stage framework for voice disorders classification. In the first stage, voice data is converted into logarithmic mel spectrograms and high-level feature embeddings are extracted from these spectrograms using the pre-trained VGGish model. In the second stage, we use classifiers, including an ensemble classifier, to classify the feature embeddings. Figure 1 provides an illustration of the proposed classification framework.

Fig. 1
figure 1

The proposed voice disorders classification system

Preprocessing and feature extraction

Before extracting features, we apply several preprocessing steps. In the original dataset, as explained in Sect. 3, voice signals were recorded at 50 kHz sampling frequency. To align with our processing requirements, we resampled the audio to 16 kHz. The length of each audio recording in the original database is different. We trimmed the audio signals to 1 s. Audio signals that were less than 1 s were padded with zeros to ensure that all audio data had the same length.

The VGGish model takes the logarithmic mel spectrogram of an audio signal for feature extraction. To compute the mel spectrogram for each audio, we apply the Short-Time Fourier Transform (STFT) with a Hamming window lasting 25 milliseconds (ms) and a 10 ms shift. This resultant spectrogram is subsequently integrated into 64 frequency bins spaced along the Mel scale, and the magnitude of each bin is then transformed logarithmically. The configuration of the mel spectrogram draws inspiration from psychoacoustic analysis, which strives to replicate characteristics of the human auditory system. This procedure involves the application of a Mel filter bank denoted as Hm(k) to filter the spectral line energy of the audio. The purpose of these filters is outlined by the following equations.

$${H_m}(k) = \left\{ {\matrix{ 0 & {{\text{if }}k < f(m - 1)} \cr {{{k - f(m - 1)} \over {f(m) - f(m - 1)}}} & {{\text{if }}f(m - 1) \le k \le f(m)} \cr {{{f(m + 1) - k} \over {f(m + 1) - f(m)}}} & {{\text{if }}f(m) < k \le f(m + 1)} \cr 0 & {{\text{if }}k > f(m + 1)} \cr } } \right.$$
(1)

Here, 0 ≤ mM, and M represents the count of filters. The central frequency f(m) of the filters can be written as:

$$f(m) = \left( {\frac{N}{{fs}}} \right)F_{mel}^{ - 1}\left( {{F_{mel}}({f_l}) + m\frac{{{F_{mel}}({f_h}) - {F_{mel}}({f_l})}}{{M + 1}}} \right)$$
(2)

Here, fl denotes the lowest frequency within the filter’s frequency domain. fh represents the highest frequency. N corresponds to the length of the Fourier transform. fs stands for the sampling frequency. Fmel signifies the Mel frequency. The transformation formula linking Fmel and the regular frequency f is given by:

$${F_{mel}} = 2595\log \left( {1 + \frac{f}{{700}}} \right)$$
(3)

The log mel spectrogram tensor (96 × 64) is the input to VGGish. Here, 96 is the number of frames within each time scale, and 64 is the number of frequency bands.

In Fig. 2, the VGGish [41] model’s structure is illustrated. Batch normalization was implemented following each convolutional layer. The chosen loss function was cross-entropy, and the model employed the Adam optimizer. Dropout, weight decay, and other usual regularization methods were not utilized. This architecture was trained on a large Audio set which was a preliminary version of the YouTube-8 M dataset. We extracted 128-dimensional high-level feature embeddings using a pre-trained VGGish model. These embeddings are then utilized as input for machine learning classifiers.

Fig. 2
figure 2

VGGish model architecture

Classifiers

We evaluated the performance of three classifiers: SVM, LR, and MLP. The SVM classifies audio signals by mapping high-dimensional VGGish features into a new space using a kernel, allowing it to create a nonlinear decision boundary. LR assigns weight coefficients to features and makes predictions based on probability scores. The MLP model learns hierarchical representations through its hidden layers, capturing complex patterns in the VGGish embeddings.

In addition, we utilized an ensemble classifier (EC) that incorporated SVM, LR, and MLP. Figure 3 shows the EC model. Instead of relying on a single model, the EC combines the predictions of multiple models to improve accuracy and reduce the risk of overfitting. Soft voting was employed to combine the predictions of these three classifiers. Soft voting is an ensemble strategy that combines the predictions of multiple classifiers by averaging their predicted probability scores for each class. In soft voting, each classifier outputs a probability distribution over the classes. For a given input sample, let pi(c) be the probability that classifier i assigns to class c. With K classifiers, the ensemble probability for class c is computed as:

$$P(c|x) = \frac{1}{K}\sum\limits_{i = 1}^K {p_i}(c)$$
(4)
Fig. 3
figure 3

Ensemble classifier

Then, the final predicted class is the one with the highest averaged probability.

$$\hat y = \arg \mathop {\max }\limits_c P(c|x) = \arg \mathop {\max }\limits_c \left( {\frac{1}{K}\sum\limits_{i = 1}^K {p_i}(c)} \right)$$
(5)

While the EC demonstrated superior performance for female speakers in the binary classification task (see Sect. 4), SVM emerged as the top-performing individual model for multiclass classification across genders and for male speakers in binary classification. This advantage can be attributed to SVM’s ability to capture complex, nonlinear relationships within high-dimensional VGGish features while maintaining robust generalization through rigorous regularization.

In our experiments, we employed SVM with a radial basis function (RBF) kernel. The SVM was configured with a regularization parameter of ‘1’, ‘scale’ kernel coefficient and utilized the ‘ovr’ (one-vs-rest) decision function shape. A logistic regression classifier with a maximum iteration count of 300, ‘newton-cg’ solver, L2 penalty and ‘ovr’ multi-class strategy was utilized to ensure convergence and prevent overfitting of the training data. Furthermore, we incorporated an MLP classifier with two hidden layers, stochastic gradient descent solver, a learning rate of 0.001 and ReLU activation function. To optimize the classifier’s performance, we employed the grid search technique. We tested all classifiers for male and female speakers separately, as well as combined for both binary and multi-class tasks.

Dataset and experimental setup

This section provides a comprehensive description of the voice dataset used in the study. We also describe the training and testing process of classifiers.

Dataset

We selected a subset of voice data from the publicly available Saarbruecken Voice Database (SVD) [42, 43] for this study. The SVD database was created by researchers at the Institut für Phonetik at Saarland University and the Phoniatry Section of the Caritas Clinic St. Theresia in Saarbrücken. The database contains audio recordings of 71 different voice disorders. Speakers engage in various speaking tasks, including the pronunciation of vowels ‘a’, ‘i’, and ‘u’ at normal, high, low, and rising-falling pitches as well as saying the sentence “Guten Morgen, wie geht es Ihnen?” (“Good morning, how are you?”). This includes individuals who were recorded before and after recovery from a voice disorder. Every recording in the database was captured at a sampling frequency of 50 kHz and 16 bits resolution.

We extracted a subset from the SVD database containing three classes: healthy, hyperfunctional dysphonia, and vocal fold paresis. The healthy class includes 227 recordings of males and 360 recordings of females. The hyperfunctional dysphonia voice disorder class has 32 recordings of males and 114 recordings of females. The vocal fold paresis voice disorder class has 25 recordings of males and 60 recordings of females. We included the recordings from individuals whose ages ranged from 19 to 60 years at the time of recording. Table 1 provides details of the subset used in this work. We chose hyperfunctional dysphonia and vocal fold paresis because they are commonly found voice disorders [44]. By choosing these specific disorders and matching the number of recordings used in previous study [45], we were able to directly compare our experimental results with existing research. This approach allowed for a more robust and fair evaluation of our findings.

Table 1 Details of voice recordings for each class

Training and testing

To train the classifiers, we used 5-fold cross-validation. In each iteration, we held out one fold for evaluation and used the remaining folds for training. All samples from each speaker were consistently placed within a single fold to prevent the model from learning to classify voice samples based on speaker identity. We computed performance metrics based on the predictions generated for the evaluation fold. The evaluation metrics include mean accuracy and F1 score, as well as mean precision, recall, and F1 score for each class.

The dataset used in this study is imbalanced, which can be a problem for machine learning models, as they can learn to favor the majority classes and ignore the minority classes. To address this issue, we balanced the training set by oversampling the minority classes. This involved duplicating samples from the minority classes to ensure that each class had an equal number of samples in the training set, which helping to prevent the model from overfitting to the majority classes. We also applied StandardScaler to all feature embeddings to ensure that all features were on the same scale, thereby to improve the performance of the machine learning models.

Experiments and results

This section presents a comprehensive overview of all the experiments conducted to compare the performance of the proposed voice disorders classification framework to other state-of-the-art methods. The first two experiments address binary classification problems: healthy vs. disordered and vocal fold paresis vs. hyperfunctional dysphonia. The third experiment is a multi-class classification problem.

Healthy vs. disordered

As shown in Table 1, we have three classes: healthy, vocal fold paresis and hyperfunctional dysphonia. For this experiment, hyperfunctional dysphonia and vocal fold paresis are combined into a single class. Separate experiments have been conducted for male speakers, female speakers, and both genders combined. Table 2 shows the mean accuracy and F1 score, as well as the precision, recall, and F1 score of each class for male and female speakers. For male speakers, VGGish-SVM achieved the highest accuracy, closely followed by VGGish-EC. VGGish-SVM achieved an accuracy of 82.45%, while VGGish-EC reached 80.25%. Our method outperforms the approach presented in [45], as shown in Table 2, which uses SVM as a classifier and features extracted with wav2vec and HuBERT models, as well as SVM with MFCC and MFCC-glottal features.

Table 2 Performance metrics for the binary classification task of healthy vs. disordered for male and female speakers

For female speakers, VGGish-EC achieved the highest accuracy with 71.54%, followed closely by VGGish-SVM, VGGish-MLP and VGGish-LR with accuracies of 70.03%, 68.36%, and 66.31% respectively. It is worth noting that this is the only case where our model demonstrates a slightly lower accuracy compared to the existing method [45], which attains its highest accuracy of 74.50% using HuBERT-SVM.

Experiments were also conducted with male and female speakers combined. The mean accuracy, F1 score, precision and recall are shown in Table 3. The results demonstrate that VGGish-EC achieved the highest overall accuracy and F1 score, with values of 73.84% and 73.92%, respectively. It was closely followed by VGGish-MLP, VGGish-SVM, and then VGGish-LR in terms of accuracy. This study’s results on combined male and female speakers can not be directly compared to those of any other study because of the differences in the datasets and disorders studied. Figure 4 presents the normalized confusion matrices for each classifier and gender.

Table 3 Performance metrics for the binary classification task of healthy vs. disordered for male and female speakers combined
Fig. 4
figure 4

Normalized confusion matrix for healthy vs. disordered. The predicted classes are represented on the horizontal axis, while the true classes are represented on the vertical axis. Class labels: 0 for healthy and 1 for disordered

Hyperfunctional dysphonia vs. vocal fold paresis

To classify hyperfunctional dysphonia and vocal fold paresis, we used the same classification setup for male, female, and combined gender speakers. The mean accuracy, F1 score, precision, and recall for male and female speakers are presented in Table 4. VGGish-SVM achieved the highest accuracy (75.45%) for male speakers, while VGGish-EC achieved 71.82%. For female speakers, VGGish-EC attained the highest accuracy at 68.42%, closely followed by VGGish-SVM, VGGish-MLP, and VGGish-LR, with respective accuracies of 68.37%, 64.97%, and 62.11%. Our method outperforms the approach presented in [45]. For male speakers, their highest accuracy was 71.95%, while for female speakers, their best accuracy was 63.06%, achieved with wav2vec-SVM.

Table 4 Performance metrics for the binary classification task of hyperfunctional dysphonia and vocal fold paresis for male and female speakers

Table 5 presents the mean accuracy, F1 score, precision, and recall for male and female speakers combined. VGGish-SVM achieved the highest overall accuracy of 68.80% and F1 score of 67.64%, followed by VGGish-EC with an accuracy of 67.10% and F1 score of 66.39%. These results cannot be directly compared to previous studies because of the differences in the datasets and disorders studied. Figure 5 illustrates the normalized confusion matrices for all classifiers and genders.

Table 5 Performance metrics for the binary classification task of hyperfunctional dysphonia and vocal fold paresis for male and female speakers combined
Fig. 5
figure 5

Normalized confusion matrix for hyperfunctional dysphonia vs. vocal fold paresis. The predicted classes are represented on the horizontal axis, while the true classes are represented on the vertical axis. Class labels: 0 for hyperfunctional dysphonia and 1 for vocal fold paresis

Multi-class classification

The mean accuracy, F1 score, precision, recall, and F1 score of each class for male and female speakers are shown in Table 6 and for both genders combined in Table 7. For male speakers, the highest classification accuracy achieved was 77.81%, for female speakers, it was 63.11%, and when both genders were combined, the accuracy reached 70.53%. In the case of multi-class classification, VGGish-SVM outperformed all other classifiers, including those in [41], in terms of accuracy. While the accuracy of VGGish-EC is lower than that of VGGish-SVM, it demonstrates better performance for the minority classes, which is important when dealing with imbalanced datasets, as it ensures that the model effectively recognizes and classifies the minority classes. The normalized confusion matrices for all classifiers are illustrates in Fig. 6. It is clear that the ensemble classifier enhances the performance of the minority classes for both male and female speakers.

Table 6 Multi-class classification performance metrics for male and female speakers
Table 7 Multi-class classification performance metrics for male and female speakers combined
Fig. 6
figure 6

Normalized confusion matrix for multi-class classification. The predicted classes are represented on the horizontal axis, while the true classes are represented on the vertical axis. Class labels: 0 for healthy, 1 for hyperfunctional dysphonia, and 2 for vocal fold paresis

Discussion

The proposed voice disorders classification system demonstrates superior performance compared to state-of-the-art methods. In this study, we employed machine learning classifiers, particularly ensemble classifiers, to evaluate high-level feature embeddings extracted using a pre-trained VGGish model. To evaluate the effectiveness of our approach, the results were compared with those reported in [45], where the same dataset was used for evaluation. Our study shows that extracting features with a pre-trained model outperforms MFCC feature-based systems, which are the most commonly used features in the detection and classification of voice disorders [46,47,48]. This statement is also confirmed by [45], where the authors extract features with the wav2vec and HuBERT models and compare the results with MFCC features.

Our study also investigated the performance of the proposed system on male and female speakers separately for both binary and multi-class classification tasks. Interestingly, our findings reveal a consistent trend where the accuracy of male speakers outperforms that of female speakers. The best accuracy for healthy vs. disordered classification was 82.45% for male speakers and 71.54% for female speakers. Similarly, the highest accuracy for hyperfunctional dysphonia vs. vocal fold paresis classification was 75.45% for male speakers and 68.42% for female speakers. In the multi-class classification scenario, the accuracy differences between male and female speakers continued similar trends. For male speakers, our model achieved an impressive accuracy of 77.81%, however, for female speakers, the highest accuracy observed was 63.11%. It is important to highlight that the binary classification of healthy vs. disordered voices for female speakers stands as the only case where our model exhibited a slightly lower accuracy compared to the results reported in [45].

VGGish-SVM achieved the highest accuracy for male speakers and VGGish-EC for female speakers in both binary classification tasks (i.e., healthy vs. disordered and hyperfunctional dysphonia vs. vocal fold paresis). In multi-class classification, VGGish-SVM performed better for both genders. However, while VGGish-EC achieved a lower overall accuracy than VGGish-SVM in multi-class classification, it outperformed VGGish-SVM on minority classes. For example, for male speakers, the precision and recall for hyperfunctional dysphonia with VGGish-SVM were 0.20% and 0.27%, respectively, while with VGGish-EC, the precision and recall were 0.23 and 0.44, respectively. Similarly, VGGish-EC performed better for vocal fold paresis. The same trend was observed for female speakers. In multi-class classification, for male speakers, the lowest F1 score is recorded for hyperfunctional dysphonia, while for female speakers, the lowest F1 score is observed for vocal fold paresis. These classes presented particular challenges in terms of accuracy, probably because of the smaller number of samples available for these classes. This underlines the importance of addressing data imbalance in future research to further enhance classification performance.

As part of our future work, we plan to incorporate explainability techniques such as LIME, SHAP, and Grad-CAM. These methods will enable us to better understand the contribution of different features in the classification process and provide visual insights into the regions of the spectrograms that are most influential in decision-making. It will help build trust in the model’s predictions and facilitate its integration into diagnostic workflows.

This study demonstrates the efficacy of hybrid frameworks for voice disorder classification using controlled datasets. However, it does not evaluate real-time performance, which is a critical factor for clinical deployment. Furthermore, the computational demands of the VGGish feature extractor and classifier pipeline may introduce latency in unoptimized implementations. Future work will focus on optimizing the framework for low-latency inference (e.g., via model lightweighting, edge-device deployment) and validating its performance on streaming audio data acquired in clinical or telehealth settings.

Conclusion

In this paper, we proposed a two-stage hybrid framework for voice disorders classification. In the first stage, we utilized a pre-trained VGGish model to extract high-level feature embeddings from the log-mel spectrograms of voice data. In the second stage, we evaluated four classifiers: support vector machine (SVM), logistic regression (LR), multilayer perceptron (MLP), and ensemble classifier.

The results of our study demonstrate the potential of using a pre-trained VGGish model to extract features for voice disorders classification. We achieved state-of-the-art results on the SVD dataset, outperforming the baseline systems that used MFCC features, MFCC-glottal features, as well as features extracted with pre-trained wav2vec and HuBERT models. Compared to the best baseline accuracy, we improved by 6.8% for male speakers in healthy vs. disordered task, 3.5% and 5.36% for male and female speakers respectively in hyperfunctional dysphonia vs. vocal fold paresis tasks. In the context of multi-class classification, our method significantly outperformed the baseline, achieving a 15.04% improvement for male speakers and a 7.75% improvement for female speakers.

While our model excelled in most scenarios, there was a slight exception. In the healthy vs. disordered task for female speakers, our model demonstrated an accuracy that was 2.96% lower when compared to the baseline. The accuracies for the combined dataset of male and female speakers are also promising in all three scenarios. It is important to note that these combined results cannot be directly compared to existing studies because of variations in the datasets and the types of voice disorders investigated.

In binary classification, VGGish-SVM exhibited the highest accuracy for male speakers, while VGGish-EC performed best for female speakers. However, in multi-class classification, VGGish-SVM outperformed other models for both genders. Notably, VGGish-EC demonstrated its strength in handling minority classes, an important aspect of medical applications. The results confirm that VGGish-EC provides more balanced accuracy by giving importance to the minority classes. Although we used oversampling to balance the classes, the accuracy of minority classes remains comparatively lower. Future research will focus on improving the robustness and generalizability of the proposed two-stage hybrid framework for voice disorders classification. Additionally, expanding the dataset to include a more diverse and a broader range of voice disorders will be crucial for enhancing the model’s applicability in real-world scenarios.

Data Availability

The data used in this study were selected from the publicly available Saarbruecken Voice Database (SVD). The full database can be accessed at the link: https://stimmdatenbank.coli.uni-saarland.de/.

References

  1. Ramig LO, Verdolini K. Treatment efficacy. J Speech Lang Hear Res. 1998;41(1):S101–S116. https://doiorg.publicaciones.saludcastillayleon.es/10.1044/jslhr.4101.s101.

    Article  CAS  PubMed  Google Scholar 

  2. Robotti C, Mozzanica F, Barillari MR, Bono M, Cacioppo G, Dimattia F, Gitto M, Rocca S, Schindler A. Treatment of relapsing functional and organic dysphonia: a narrative literature review. Acta Otorhinolaryngol Ital. 2023;43(2 Suppl 1):84. https://doiorg.publicaciones.saludcastillayleon.es/10.14639/0392-100x-suppl.1-43-2023-11.

    Article  Google Scholar 

  3. American Speech-Language-Hearing Association. (n.d.).Voice disorders. (Practice Portal). Accessed 14 Sept 2023. https://www.asha.org/practice-portal/clinical-topics/voice-disorders/

  4. Ribas D, Pastor MA, Miguel A, Martnez D, Ortega A, Lleida E. Automatic voice disorder detection using self-supervised representations. IEE Access. 2023;11:14915–27. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ACCESS.2023.3243986.

    Article  Google Scholar 

  5. Xie Y, Ruiyu L, Liang Z, Huang C, Zou C, Schuller B. Speech emotion classification using attention-based LSTM. IEEE/ACM Trans Audio Speech Lang Process. 2019;27:1–1. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/TASLP.2019.2925934.

    Article  Google Scholar 

  6. Keser S, Gezer E. Comparative analysis of speaker identification performance using deep learning, machine learning, and novel subspace classifiers with multiple feature extraction techniques. Digital Signal Process. 2025;156:104811. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.dsp.2024.104811.

    Article  Google Scholar 

  7. Korkmaz Y, Boyacı A. milVAD: a bag-level mnist modelling of voice activity detection using deep multiple instance learning. Biomed Signal Process Control. 2022;74:103520. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.bspc.2022.103520.

    Article  Google Scholar 

  8. Fang S-H, Tsao Y, Hsiao M-J, Chen J-Y, Lai Y-H, Lin F-C, Wang C-T. Detection of pathological voice using cepstrum vectors: a deep learning approach. J Voice. 2019;33(5):634–41. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jvoice.2018.02.003.

    Article  PubMed  Google Scholar 

  9. Cordeiro H, Fonseca J, Guimarães I, Meneses C. Hierarchical classification and system combination for automatically identifying physiological and neuromuscular laryngeal pathologies. J Voice. 2017;31(3):384.e9–384.e14. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jvoice.2016.09.003.

    Article  PubMed  Google Scholar 

  10. Kodrasi I, Pernon M, Laganaro M, Bourlard H. Automatic and perceptual discrimination between dysarthria, apraxia of speech, and neurotypical speech. In: ICASSP 2021–2021 IEEE international conference on acoustics, speech and signal processing (ICASSP). 2021. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/icassp39728.2021.9414283

  11. Costa SC, Neto BGA, Fechine JM. Pathological voice discrimination using CEPSTRAL analysis, vector quantization and Hidden Markov models. In: 2008 8th IEEE international conference on bioinformatics and bioengineering. 2008. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/bibe.2008.4696783.

  12. Salhi L, Mourad T, Cherif A. Voice disorders classification using multilayer neural network. In: 2008 2nd International conference on signals, circuits and systems. 2008. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/icscs.2008.4746953.

  13. Fraile R, Sáenz-Lechón N, Godino-Llorente JI, Osma-Ruiz V, Fredouille C. Automatic detection of laryngeal pathologies in records of sustained vowels by means of Mel-frequency cepstral coefficient parameters and differentiation of patients by sex. Folia Phoniatr Logop. 2009;61(3):146–52. https://doiorg.publicaciones.saludcastillayleon.es/10.1159/000219950.

    Article  CAS  PubMed  Google Scholar 

  14. Ali Z, Elamvazuthi I, Alsulaiman M, Muhammad G. Automatic voice pathology detection with running speech by using estimation of auditory spectrum and CEPSTRAL coefficients based on the all-pole model. J Voice. 2016;30(6):757.e7–757.e19. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jvoice.2015.08.010.

    Article  PubMed  Google Scholar 

  15. Al-Dhief FT, Latiff NMA, Malik NNNA, Baki MM, Sabri N, Albadr MAA. Dysphonia detection based on voice signals using naive bayes classifier. In: 2022 IEEE 6th international symposium on telecommunication technologies (ISTT). IEEE; 2022. pp. 56–61. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/istt56288.2022.9966535.

  16. Benba A, Jilbab A, Hammouch A. Analysis of multiple types of voice recordings in cepstral domain using MFCC for discriminating between patients with Parkinson’s disease and healthy people. Int J Speech Technol. 2016;19(3):449–56. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s10772-016-9338-4.

    Article  Google Scholar 

  17. Gómez-García JA, Moro-Velázquez L, Godino-Llorente JI. On the design of automatic voice condition analysis systems. Part ii: review of speaker recognition techniques and study on the effects of different variability factors. Biomed Signal Process Control. 2019;48:128–43. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.bspc.2018.09.003.

    Article  Google Scholar 

  18. Reddy MK, Alku P. A comparison of CEPSTRAL features in the detection of pathological voices by varying the input and Filterbank of the cepstrum computation. IEEE Access. 2021;9:135953–63. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/access.2021.3117665.

    Article  Google Scholar 

  19. Yagnavajjula MK, Mittapalle KR, Alku P, Rao KS, Mitra P. Automatic classification of neurological voice disorders using wavelet scattering features. Speech Commun. 2024;157:103040. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.specom.2024.103040.

    Article  Google Scholar 

  20. Martínez D, Lleida E, Ortega A, Miguel A, Villalba J. Voice pathology detection on the Saarbrücken voice database with calibration and fusion of scores using multifocal toolkit. In: Advances in speech and language technologies for Iberian languages. 2012. pp. 99–109. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-3-642-35292-8_11.

  21. Parsa V, Jamieson DG. Acoustic discrimination of pathological voice. J Speech Lang Hear Res. 2001;44(2):327–39. https://doiorg.publicaciones.saludcastillayleon.es/10.1044/1092-4388(2001/027).

    Article  CAS  PubMed  Google Scholar 

  22. Fujimura S, Kojima T, Okanoue Y, Shoji K, Inoue M, Omori K, Hori R. Classification of voice disorders using a one-dimensional convolutional neural network. J Voice. 2022;36(1):15–20. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jvoice.2020.02.009.

    Article  PubMed  Google Scholar 

  23. Wu H, Soraghan J, Lowit A, Di-Caterina G. A deep learning method for pathological voice detection using convolutional deep belief networks. In: Interspeech 2018. 2018. https://doiorg.publicaciones.saludcastillayleon.es/10.21437/interspeech.2018-1351.

  24. Wu H, Soraghan J, Lowit A, Caterina GD. Convolutional neural networks for pathological voice detection. In: 2018 40th Annual international conference of the IEEE engineering in Medicine and Biology Society (EMBC). 2018. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/embc.2018.8513222.

  25. Mohammed MA, Abdulkareem KH, Mostafa SA, Khanapi Abd Ghani M, Maashi MS, Garcia-Zapirain B, Oleagordia I, Alhakami H, Al-Dhief FT. Voice pathology detection and classification using convolutional neural network model. Appl Sci. 2020;10(11):3723. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/app10113723.

    Article  CAS  Google Scholar 

  26. Islam R, Tarique M. A novel convolutional neural network based dysphonic voice detection algorithm using Chromagram. Int J Elec Comp Eng (IJECE). 2022;12(5):5511–18. https://doiorg.publicaciones.saludcastillayleon.es/10.11591/ijece.v12i5.pp5511-5518.

    Article  Google Scholar 

  27. Chaiani M, Selouani SA, Boudraa M, Yakoub MS. Voice disorder classification using speech enhancement and deep learning models. Biocybern Biomed Eng. 2022;42(2):463–80. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.bbe.2022.03.002.

    Article  Google Scholar 

  28. Harar P, Alonso-Hernandezy JB, Mekyska J, Galaz Z, Burget R, Smekal Z. Voice pathology detection using deep learning: a preliminary study. In: 2017 International conference and workshop on bioinspired intelligence (IWOBI). IEEE; 2017. pp. 1–4. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/iwobi.2017.7985525.

  29. Korkmaz Y. SS-ESC: a spectral subtraction denoising based deep network model on environmental sound classification. Signal Image Video Process. 2024;19:50. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s11760-024-03649-5.

    Article  Google Scholar 

  30. Korkmaz Y, Boyacı A. Classification of Turkish vowels based on formant frequencies. In: 2018 International conference on artificial intelligence and data processing (IDAP). 2018. pp. 1–4. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/IDAP.2018.8620877.

  31. Hernandez A, Pérez-Toro PA, Nöth E, Orozco-Arroyave JR, Maier A, Yang SH. Cross-lingual self-supervised speech representations for improved dysarthric speech recognition. arXiv Preprint arXiv:2204 01670. 2022. https://doiorg.publicaciones.saludcastillayleon.es/10.21437/interspeech.2022-10674.

  32. Rahman MU, Direkoglu C. Multi-class classification of voice disorders using deep transfer learning. In: Computing, internet of things and data analytics. ICCIDA 2023. Studies in computational intelligence. Studies in computational intelligence, vol. 1145. Cham, Switzerland: Springer; 2024. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-3-031-53717-2_25.

  33. Mallela J, Illa A, Suhas B, Udupa S, Belur Y, Atchayaram N, Yadav R, Reddy P, Gope D, Ghosh PK. Voice based classification of patients with amyotrophic lateral sclerosis, Parkinson’s disease and healthy controls with CNN-LSTM using transfer learning. In: ICASSP 2020–2020 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE. 2020. pp. 6784–88. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/icassp40776.2020.9053682.

  34. Zaman K, Sah M, Direkoglu C, Unoki M. A survey of audio classification using deep learning. IEEE Access. 2023;11:106620–49. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ACCESS.2023.3318015.

    Article  Google Scholar 

  35. Peng X, Xu H, Liu J, Wang J, He C. Voice disorder classification using convolutional neural network based on deep transfer learning. Sci Rep. 2023;13(1):7264. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41598-023-34461-9.

  36. Violeta LP, Huang WC, Toda T. Investigating self-supervised pretraining frameworks for pathological speech recognition. In: Interspeech 2022. 2022. https://doiorg.publicaciones.saludcastillayleon.es/10.21437/interspeech.2022-10043.

  37. Zhu Y, Liang X, Batsis JA, Roth RM. Domain-aware intermediate pretraining for dementia detection with limited data. In: Interspeech 2022. 2022. https://doiorg.publicaciones.saludcastillayleon.es/10.21437/interspeech.2022-10862.

  38. Zhu Y, Obyat A, Liang X, Batsis JA, Roth RM. Wavbert: exploiting semantic and non-semantic speech using Wav2vec and BERT for dementia detection. In: Interspeech 2021. 2021. https://doiorg.publicaciones.saludcastillayleon.es/10.21437/interspeech.2021-332.

  39. Karaman O, Çakın H, Alhudhaif A, Polat K. Robust automated Parkinson disease detection based on voice signals with transfer learning. Expert Syst Appl. 2021;178:115013. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.eswa.2021.115013.

    Article  Google Scholar 

  40. Sekhar SRM, Kashyap G, Bhansali A, Abishek A A, Singh K. Dysarthric-speech detection using transfer learning with convolutional neural networks. ICT Express. 2022;8(1):61–64. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.icte.2021.07.004.

    Article  Google Scholar 

  41. Hershey S, Chaudhuri S, Ellis DP, Gemmeke JF, Jansen A, Moore RC, Plakal M, Platt D, Saurous RA, Seybold B, et al. CNN architectures for large-scale audio classification. In: 2017 Ieee International conference on acoustics, speech and signal processing (icassp). IEEE; 2017. pp. 131–35. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/icassp.2017.7952132.

  42. Pützer M, Barry WJ. Saarbrücken Voice Database. Institute of Phonetics, University of Saarland. http://www.stimmdatenbank.coli.uni-saarland.de/.

  43. Pützer M, Barry WJ. Instrumental dimensioning of normal and pathological phonation using acoustic measurements. Clin Linguist Phon. 2008;22(6):407–20. https://doiorg.publicaciones.saludcastillayleon.es/10.1080/02699200701830869.

    Article  PubMed  Google Scholar 

  44. Walton C, Conway E, Blackshaw H, Carding P. Unilateral vocal fold paralysis: a systematic review of speech-language pathology management. J Voice. 2017;31(4):509.e7–509.e22. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jvoice.2016.11.002.

    Article  PubMed  Google Scholar 

  45. Tirronen S, Kadiri SR, Alku P. Hierarchical multi-class classification of voice disorders using self-supervised models and glottal features. IEEE Open J Signal Process. 2023;4:80–88. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ojsp.2023.3242862.

    Article  Google Scholar 

  46. Amara F, Fezari M, Bourouba H. An improved GMM-SVM system based on distance metric for voice pathology detection. Appl Math Inf Sci. 2016;10(3):1061–70. https://doiorg.publicaciones.saludcastillayleon.es/10.18576/amis/100324.

    Article  Google Scholar 

  47. Laguarta J, Hueto F, Subirana B. Covid-19 artificial intelligence diagnosis using only cough recordings. IEEE Open J Eng Med Biol. 2020;1:275–81. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ojemb.2020.3026928.

    Article  PubMed  Google Scholar 

  48. Rejaibi E, Komaty A, Meriaudeau F, Agrebi S, Othmani A. MFCC-based recurrent neural network for automatic clinical depression recognition and assessment from speech. Biomed Signal Process Control. 2022;71:103107. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.bspc.2021.103107.

    Article  Google Scholar 

Download references

Acknowledgements

Not applicable.

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Contributions

Concept and design: MUR. Experiments and analysis: MUR. Drafting of the manuscript: MUR and CD. Critical revision of the manuscript: MUR and CD.

Corresponding author

Correspondence to Mehtab Ur Rahman.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Clinical trial number

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Rahman, M., Direkoglu, C. A hybrid approach for binary and multi-class classification of voice disorders using a pre-trained model and ensemble classifiers. BMC Med Inform Decis Mak 25, 177 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12911-025-02978-w

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12911-025-02978-w

Keywords