Next Article in Journal
A Histogram Publishing Method under Differential Privacy That Involves Balancing Small-Bin Availability First
Previous Article in Journal
Prime Time Tactics—Sieve Tweaks and Boosters
Previous Article in Special Issue
Ensemble Learning with Pre-Trained Transformers for Crash Severity Classification: A Deep NLP Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Central Kurdish Text-to-Speech Synthesis with Novel End-to-End Transformer Training

by
Hawraz A. Ahmad
1 and
Tarik A. Rashid
2,*
1
Department of Software and Informatics Engineering, Salahaddin University-Erbil, Erbil 44001, Iraq
2
Department of Computer Science and Engineering, University of Kurdistan Hawler, Erbil 44001, Iraq
*
Author to whom correspondence should be addressed.
Algorithms 2024, 17(7), 292; https://doi.org/10.3390/a17070292
Submission received: 30 April 2024 / Revised: 18 June 2024 / Accepted: 2 July 2024 / Published: 3 July 2024
(This article belongs to the Special Issue AI Algorithms for Positive Change in Digital Futures)

Abstract

:
Recent advancements in text-to-speech (TTS) models have aimed to streamline the two-stage process into a single-stage training approach. However, many single-stage models still lag behind in audio quality, particularly when handling Kurdish text and speech. There is a critical need to enhance text-to-speech conversion for the Kurdish language, particularly for the Sorani dialect, which has been relatively neglected and is underrepresented in recent text-to-speech advancements. This study introduces an end-to-end TTS model for efficiently generating high-quality Kurdish audio. The proposed method leverages a variational autoencoder (VAE) that is pre-trained for audio waveform reconstruction and is augmented by adversarial training. This involves aligning the prior distribution established by the pre-trained encoder with the posterior distribution of the text encoder within latent variables. Additionally, a stochastic duration predictor is incorporated to imbue synthesized Kurdish speech with diverse rhythms. By aligning latent distributions and integrating the stochastic duration predictor, the proposed method facilitates the real-time generation of natural Kurdish speech audio, offering flexibility in pitches and rhythms. Empirical evaluation via the mean opinion score (MOS) on a custom dataset confirms the superior performance of our approach (MOS of 3.94) compared with that of a one-stage system and other two-staged systems as assessed through a subjective human evaluation.

1. Introduction

In recent years, text-to-speech (TTS) systems have undergone significant advancements, which were largely driven by the adoption of deep learning (DL) techniques. These systems transform written text into natural-sounding speech through a multi-stage process. However, traditional approaches often necessitate trade-offs between the quality of the synthesized speech, the speed of generation, and the complexity of the training process.
One prevalent approach leverages a two-stage architecture [1,2]. In the initial stage, the system generates intermediate representations, such as linguistic features [2] or mel-spectrograms—a representation capturing the frequency content of an audio signal over time [1] from the preprocessed text. To generate mel-spectrograms from the preprocessed text, the system typically utilizes techniques like text-to-speech (TTS) models that convert the text into spectrogram images. These spectrograms represent how the frequency content of the audio signal varies over time, providing a structured format for subsequent stages of audio synthesis. The second stage then translates these representations into raw audio waveforms [2,3]. While this method has yielded realistic speech, it suffers from limitations. Training these models often requires sequential training or fine-tuning, hindering efficiency. Additionally, their reliance on predefined intermediate features restricts the utilization of potentially beneficial learned representations, limiting the system’s ability to further improve performance.
To address these shortcomings, alternative approaches have been explored, such as non-autoregressive methods [4,5] and generative adversarial networks (GANs) [6]. Non-autoregressive models aim to overcome the slow generation speed inherent to autoregressive systems such as Tacotron 2 [1] and Transformer TTS [7] by eliminating the sequential nature of the process. These models can synthesize speech significantly faster, making them more suitable for real-time applications. On the other hand, GAN-based methods have shown promise in generating high-quality waveforms, potentially surpassing the quality achieved by two-stage approaches [8,9]. Recent efforts have focused on developing efficient end-to-end training methods for TTS models [10,11]. These methods aim to bypass the two-stage pipeline entirely and directly convert text into speech. While these approaches offer potential performance improvements by leveraging learned representations throughout the entire process, they often fall short of the quality achieved by two-stage systems, highlighting the ongoing challenge of balancing efficiency and quality in TTS.
The Sorani Kurdish dialect is predominantly spoken by Kurdish communities in Iraq and Iran, representing a significant linguistic and cultural identity within these regions. Despite its importance, Sorani Kurdish has faced challenges in modern technological advancements, particularly in the realm of TTS conversion. TTS technology plays a crucial role in enhancing accessibility and inclusivity for languages and dialects worldwide. However, the development of TTS systems for Sorani Kurdish has been relatively limited compared to more widely spoken languages. This discrepancy poses a barrier to the full participation of Sorani Kurdish speakers in digital communication, education, and accessibility tools that rely on TTS technology.
This study presents a parallel end-to-end TTS method to address the limitations of both traditional two-stage architectures and recent end-to-end methods. The proposed model is named Kurdish TTS (KTTS), as it will be developed to generate more natural-sounding Central Kurdish audio than that generated by current two-stage models. The goal of this study is to achieve high-quality speech synthesis while maintaining efficiency and simplified training protocols. Our approach utilizes a variational autoencoder (VAE) that is pre-trained for audio waveform reconstruction. This involves aligning the prior distribution established by the pre-trained encoder with the posterior distribution of the text encoder within latent variables. To augment the expressive capabilities of our method and enable the synthesis of high-fidelity speech waveforms, we employ adversarial training [6] in the waveform reconstruction. By aligning latent distributions and integrating the stochastic duration predictor, our method facilitates the real-time generation of natural Kurdish audio speech. The proposed model is trained directly to maximize the log-likelihood of speech, and this is coupled with the alignment process.
One of the crucial steps in developing a TTS system is the creation of a high-quality speech corpus. Developing models that capture the prosodic patterns of Kurdish is essential for creating natural-sounding synthesized speech. Other contributions of this research work can be summarized as follows:
  • A novel end-to-end method for Kurdish text-to-speech conversion based on a VAE framework is introduced. The proposed VAE effectively maps input waveforms to a latent space representation and reconstructs them.
  • A robust training procedure is developed to align the latent variables of the text encoder with those of the pre-trained waveform encoder of the VAE. This involves ensuring that the prior distribution established by the pre-trained encoder matches with the posterior distribution of the text encoder within latent variables.
  • The proposed KTTS directly regenerates waveforms from input text by bypassing the intermediate stages required to create mel-spectrograms or linguistic features.
  • A new dataset comprising aligned pairs of Central Kurdish text sequences and corresponding audio recordings is curated. This dataset serves as a valuable resource for advancing research in Kurdish text-to-speech synthesis.
The rest of this article is organized as follows: Section 2 offers a review of the TTS literature, Section 3 explains the methodology for KTTS, and Section 4 details the experimental setup, including the dataset and training parameters. In Section 5, we present the results and a discussion, and finally, Section 6 provides conclusions based on our findings.

2. TTS Literature Review

This section provides an overview of existing TTS systems, categorizing them into one-stage and two-stage approaches. It also briefly discusses the existing literature on Kurdish TTS systems. Transformers, initially introduced by Vaswani et al. in 2017 [12], have revolutionized natural language processing by capturing long-range dependencies among input tokens, which is particularly beneficial for tasks like text-to-speech (TTS) synthesis. In recent years, their application to TTS development has yielded significant improvements in the naturalness and intelligibility of synthesized speech. This advance underscores the versatility and power of transformers in handling complex sequential data, highlighting their potential in other domains as well. In the following two subsections, we summarize state-of-the-art TTS systems, which are mostly based on transformer architectures.

2.1. One-Stage Systems

One-stage text-to-speech (TTS) systems streamline the process of converting text into synthetic speech by employing a direct transformation model. These systems leverage end-to-end neural network architectures, such as sequence-to-sequence models with attention mechanisms, to map raw text directly to acoustic waveforms. In variational inference with adversarial learning for end-to-end text-to-speech (VITS) [13], a duration predictor is introduced to improve the rhythm of the generated utterances. VITS was extended to allow the generation of diverse utterances for multi-language speakers using your-TTS. Although these models allow sampling from the input tokens, the quality of these generated utterances is still inferior to that obtained with single-speaker systems [14].
In ref. [4], the authors introduced FastSpeech, a non-autoregressive version of transformer TTS [7]. They used the original model as a teacher and extracted the character durations from it. To generate all output frames, they trained a student model using a convolutional duration prediction.
In 2021, Ren et al. introduced Fast Speech 2 [10], a non-autoregressive version of transformer TTS. The researchers used external durations to improve the training process and reduce the development costs. This approach assumes that the alignment model used for the language is of high quality.
Recent advancements include models like VITS 2 [15], which combines variational inference with normalizing flows and adversarial learning to directly generate high-fidelity speech waveforms from text. In ref. [16], the authors proposed a framework for building controllable TTS systems that can generate speech with specific attributes. It combines a sequence-to-sequence TTS model with a conditional variational autoencoder (CVAE) to learn disentangled representations of speech attributes. The system enables flexible and controllable speech synthesis. The integration of large language models, such as LLaMA, into TTS systems, has been shown to enhance semantic understanding and generate more expressive speech [17], highlighting the potential for semantic-aware TTS systems to further improve synthesis quality.
Our proposed model lies in the category of one-stage systems, where no intermediate stages are needed to create mel-spectrograms and then to convert mel-spectrograms into waveforms, as our system regenerates waveforms directly from input text.

2.2. Two-Stage Systems

Two-stage TTS systems introduce an intermediate step between text processing and waveform generation, typically predicting a mel-spectrogram before synthesizing the final speech output. This approach separates the linguistic and acoustic modeling stages, allowing for more fine-grained control and potentially higher-quality synthetic speech.
The SC-GlowTTS system [18] is a flow-based multi-speaker text recognition system that takes the predicted parameters of an external speaker embedding into account. SNAC [19], on the other hand, utilizes a coupling layer to explicitly normalize the input.
The basic Glow-TTS [20] architecture consists of a flow-based determinate duration predictor, a transformer-equipped encoder, and a flow-dependent decoder. The transformer-based encoder produces a linear approximation of the prior distribution mean by translating the input tokens’ phonetic embedding into a representation with an 80-dimensional structure. The z-sampling method can also be utilized to express the distribution’s z-sampled value:
z = μ + T ϵ ,
In training, the duration predictor only predicts the mean µ and the temperature T, while at inference time, it chooses a value of T that is usually smaller than 1. A latent representation of the distribution is then sampled from the prior data to generate a mel-spectrogram.
In 2017, Vaswani et al. introduced the concept of transformer TTS [12]. In 2019, Li et al. [7] tested the effectiveness of this technology by developing an algorithm that can predict the mel-spectrogram for English phonemes. The evaluation of the transformer TTS system showed that it was very promising, but it was not feasible to use it in a production setting because the auto-regressive approach was time-consuming. The evaluation of the mean opinion score (MOS) of a phonemicized dataset using a non-autoregressive model was not significantly different from the results when using the transformer TTS system. The authors also used a pitch prediction module with FastPitch [21] to complement the duration predictor in their work from the Tactron2 model that they introduced [1]. The authors claimed that the quality of their results was similar when using durations and phonemes from a Montreal forced alignment (MFA) model [22].
Diff-TTS [23] uses a diffusion probabilistic model to first generate mel-spectrograms, which are then converted to speech using advanced vocoder models like HiFi-GAN [9]. More recently, models such as WaveGrad 2 [24] and EfficientTTS 2 [25] focus on optimizing the two-stage process for faster and more efficient synthesis without compromising on quality. In 2023, the MelStyleTTS [26] proposed a style transfer technique for mel-spectrograms, allowing for greater expressiveness in synthetic speech. Two-stage systems have been shown to produce more natural and expressive speech compared to their one-stage counterparts, although they may introduce additional latency and complexity in the synthesis pipeline.

3. Related Work for the Kurdish Language

After conducting an extensive review of the existing literature on the Kurdish language, it is apparent that the majority of previous research has focused on utilizing and adapting existing models rather than developing novel approaches. We have summarized some of the works in Table 1, which shows some initial works in Kurdish and some Kurdish TTS approaches.

4. Methodology

This section explains the proposed method and its architecture. As illustrated in Figure 1, our approach for Kurdish text-to-speech conversion comprises three key procedures: VAE for waveform reconstruction (Figure 1A), training (Figure 1B), and inference (Figure 1A). Detailed descriptions of the components and blocks employed within our framework will be explained in this section.

4.1. Variational Autoencoder

To effectively pre-train a VAE for waveform reconstruction, several key components and formulas are essential. The VAE framework aims to learn a probabilistic mapping from an input waveform x to a latent space representation z and subsequently reconstructs the input as x ^ . This process involves two primary objectives: maximizing the likelihood of generating the input data given the latent variables and enforcing the learned latent space to follow a prior distribution.
The wave encoder q ϕ ( z | x ) , an approximate posterior distribution that is parameterized by ϕ , maps the input waveform x to a latent space representation z, where z N ( μ ϕ , σ ϕ 2 ) , with μ ϕ and σ ϕ 2 representing the mean and variance of the latent space distribution, respectively. Subsequently, the wave decoder p ψ ( x ^ | z ) parameterized by ψ generates the reconstructed output x ^ conditioned on the sampled latent variable z.
For training our VAE, the total loss L ( ψ , ϕ ) is the sum of two terms: the reconstruction loss L r e c and Kullback–Leibler ( K L ) divergence.
L ( ψ , ϕ ) = E z q ϕ ( z | x ) [ log p ψ ( x | z ) ] + D K L [ q ϕ ( z | x ) | | p ψ ( z ) ]
L r e c is expressed by the negative log-likelihood log p ψ ( x | z ) capturing the probabilistic aspect of reconstruction, ensuring that the generated output closely resembles the input data distribution. For L r e c , we simply use the mean square error (MSE) between the input x and the reconstructed output x ^ . Second, the K L divergence term enforces a regularization constraint, guiding the latent space towards a predefined prior distribution p ψ ( z ) , which is a Gaussian distribution of N ( 0 , 1 ) . Now, the target is to find the optimal ψ and ϕ such that
ψ * , ϕ * = arg max ψ , ϕ L ( ψ , ϕ )

4.2. Training Procedure

During the training, we take the pre-trained wave encoder q ϕ ( z | x ) from the previous step with its parameters ϕ frozen to generate the latent representation z of the target waveform x.
Our ultimate goal is to model the conditional distribution of the waveform data p θ , A , ψ ( x | c ) by transforming a conditional prior distribution p θ , A ( z | c ) through the pre-trained wave decoder p ψ ( x | z ) , where c represents the input text sequence (see Figure 1C). We parameterize the prior distribution with the parameters θ of the text encoder and an alignment function A, which is discussed in Section 4.3.
To achieve this, we need to minimize the distance between the posterior distribution of the pre-trained wave encoder p ϕ ( z | x ) and the prior distribution p θ , A ( z | c ) . Once again, we employ KL divergence to force the latent space z to conform to p ϕ ( z | x ) . The KL divergence is, then,
D K L = log q ϕ ( z | x ) log p θ , A ( z | c )
where
z q ϕ ( z | x ) = N ( z ; μ ϕ ( x ) , σ ϕ ( x ) )
The prior distribution’s statistics, denoted as μ θ and σ θ , are computed using the text encoder, which can transform the text condition c = c 1 : T t e x t into the corresponding statistics, μ = μ 1 : T t e x t and σ = σ 1 : T t e x t , with T t e x t representing the length of the input text.

4.3. Alignment Prediction

The alignment function A denotes the mapping from the index of the latent representation of waveform z to the corresponding index of statistics from the text encoder, A ( j ) = i , whenever z j follows a normal distribution N ( z j ; μ i , σ i ) . We presume that A maintains both monotonicity and subjectivity to prevent skipping or repeating the input text. Subsequently, the prior distribution can be articulated as follows:
log p θ , A ( z | c ) = j T w a v e f o r m log N ( z j ; μ A ( j ) , σ A ( j ) ) ,
where T w a v e f o r m is the length of the input waveform.
Similar to [20], we employ a monotonic alignment search to find the parameters θ and the alignment A that maximize the log-likelihood of waveform data, as shown in Equation (7).
max θ , A L ( θ , A ) = max θ , A p θ , A , ψ ( x | c )
Throughout the training process, we keep the parameters of the pre-trained VAE ϕ and ψ frozen. Consequently, our objective is to find the optimal alignment function A * , after which we update θ using gradient descent:
A * = arg max A log p θ , A , ψ ( x | c ) = arg max A j T w a v e f o r m log N ( z j ; μ A ( j ) , σ A ( j ) )

4.4. Duration Prediction

Given the absence of ground-truth labels for the alignment, it becomes necessary to estimate the alignment at every training iteration. The duration of each input token d i can be computed by summing the columns within each row of the estimated alignment, as shown in Equation (9). This duration calculation serves as our ground truth for training a deterministic duration predictor, f d u r a t i o n .
d i = j T w a v e f o r m 1 A * ( j ) = i , i = 1 , 2 , . . . . . T t e x t
During the training procedure, we train f d u r a t i o n to re-predict the duration computed in Equation (9) from the optimal alignment A * . This duration prediction also helps predict A * during the inference process. We train f d u r a t i o n with the MSE loss, as outlined in Equation (10), by integrating it on top of the text encoder (see Figure 1). In order to prevent interference with the maximum likelihood objective, we employ the stop gradient technique on the input of the duration predictor during the backward pass [33].
L d u r a t i o n = M S E ( D , d ) ,
where
D = f d u r a t i o n ( S G ( h ) )
where S G denotes the stop gradient operator, and h t e x t is the hidden representation of the text encoder.

4.5. Inference Procedure

Throughout the inference process, which is illustrated in Figure 1C, the statistical parameters μ θ and σ θ of the prior distribution, along with A * , are obtained by the text encoder and duration predictor. Then, a latent variable is sampled from the prior distribution z N ( μ θ , σ θ 2 ) , and concurrently, a waveform x ^ is synthesized by transforming the sampled z using the pre-trained wave decoder. Instead of feeding the entire latent representation z, we segment z into slices with a size of 32, each corresponding to a brief audio clip. The pre-trained wave decoder sequentially receives the slices and up-samples (transforms) them to the corresponding audio clips.

4.6. Model Architecture

4.6.1. Text Encoder

To handle Central Kurdish text, our initial step involves converting text sequences into International Phonetic Alphabet (IPA) sequences through the utilization of open-source software [34]. Additionally, we incorporate several custom-defined phonemes to accommodate the distinct characters present in Central Kurdish, as outlined in Appendix A.2. Then, the text encoder converts the phoneme embedding sequence into the hidden phoneme representation h t e x t . We follow the encoder structure of the transformer [12], as shown in Figure 2, with some slight modifications. We remove the positional encoding and add learnable positional encoding. We build the text encoder with eight blocks of transformer encoders, each with eight multi-head self-attention modules. The dimension of phoneme embeddings and the hidden size of the self-attention (hidden representations) are set to 256 following the recommendation by FASTSPEECH [10].
The positional encoder depicted in Figure 3 employs a grouped 1D convolution comprising 64 filters with a kernel size of 3 to generate a relative positional vector from the latent features. This vector is subsequently combined with the embedding of phonemes (tokens) to encode their positions relative to each other. We append a linear projection layer on top of the transformer encoder to predict the statistics of the prior distribution, μ θ and σ θ , from the hidden representation h t e x t .

4.6.2. Wave Encoder

To build our wave encoder, we utilize a transformer encoder structure identical to that employed in the text encoder, as depicted in Figure 4. This choice aids the text encoder network in converging more swiftly when the KL divergence is applied for the difference between the two distributions.
In designing the feature encoding block, we adopt a similar structure to that outlined in ref. [35] with slight changes, as depicted in Figure 5. To enhance the processing efficiency, we opted for a configuration of five 1D convolutional blocks instead of the original seven, achieving comparable results according to empirical validation. Additionally, we substitute the GEUL activation layers with PRELU. The receptive field of the feature encoder spans a total context of 2200 samples, corresponding to 100 ms at the 22 kHz input sample rate. Consequently, this feature encoder extracts features from the raw waveform and tokenizes it, with each token representing a 100 ms segment. These tokens are subsequently processed by the transformer encoder.

4.6.3. Wave Decoder

Our wave decoder architecture is modeled after WaveNet [2], as illustrated in Figure 6. It comprises a transposed 1D convolution with a filter size of 64, along with 30 dilated residual convolution blocks. The skip channel size and kernel size of the 1D convolution are configured to 64 and 3, respectively. The wave decoder receives a sliced hidden representation z with a channel size of 256 generated by the wave or text encoder, corresponding to a brief audio clip, as its input. It then utilizes transposed 1D convolution to upsample the slice, aligning it with the length of the corresponding audio clip.
Similar to prior works [2,13,36], we incorporate adversarial training into the wave decoder. The discriminator D in the adversarial training adopts the same structure and configurations as those of Parallel WaveGAN [36]. D distinguishes between the waveform x ^ generated by the wave decoder and the ground-truth waveform x. We optimize the wave decoder by incorporating the multi-resolution short-time Fourier transform (STFT) loss in conjunction with the discriminator loss from the least squares generative adversarial network (LSGAN), aligning with the methodology of Parallel WaveGAN [36].

4.6.4. Duration Predictor

f d u r a t i o n predicts the distribution of phoneme durations from the hidden representation h t e x t . To build its architecture, we stack two residual blocks, as shown in Figure 7. Each of these blocks consists of a convolutional layer containing 256 filters, each with a kernel size of 3, alongside a PRELU activation function and layer normalization followed by an FC. PRELU is chosen for its ability to learn negative slope values, mitigating the issue of dead neurons associated with RELU. Additionally, the inclusion of residual connections serves to mitigate vanishing gradients, thereby enhancing performance and reducing overfitting by encouraging feature reuse.

5. Experiments

This section details the dataset creation, categorization, and recording process, followed by an overview of the corpus statistics and technical specifications. The training approach, including the dataset partitioning and pre-training of the VAE, is outlined, along with the GPU usage and optimization techniques, providing a concise overview of our experimental setup.

5.1. Dataset

Text-to-speech systems depend on the availability of a corpus containing pairs of speech and corresponding text. This study explores voice data of the Central Kurdish dialect for TTS systems. We started by creating an audio- and text-pairing dataset featuring a male individual who spoke in Central Kurdish. The recording process was carried out by a male dubber in a recording studio. The 6078 sentences that we collected from the text corpus were categorized into 12 categories, including sports, science, literature, health, and everything else. Training sentences were then created using the collected information, resulting in 4255 (70%) sentences. The validation set contained 608 (10%) random sentences. The testing set contained 1215 (20%) sentences that were randomly selected from the overall dataset. The sentences were then improved through various web sources. Table 2 illustrates the subjects and the number of sentences. The process of recording speech files ended after 30 days. The dataset can be accessed through the following link [37].
Some features of the recorded files can be summarized as follows: (a) 6078 WAV files and over 13.63 h of recorded speech were captured; (b) the output of the files was recorded at a rate of 22,050 kHz; (c) the quantization process was carried out using 16 bits of signed data; (d) the stored speech audio files were in the format known as PCM, and a mono channel was utilized to record the audio streams; (e) the shortest audio file length was 0.502 s; (f) the longest audio file length was 16.781 s; (g) the mean audio file length was 8.076 s.
The audio files are stored in wave format, while the text sentences are saved in an Excel file. The audio files are organized in a single folder. The audio file’s name includes the extension names, while the transcript is the text of the speech referenced to the audio file with an ID which is the name of the audio file. The dataset was prepared to comply with the Gaussian distribution to be more effective in training models avoiding bias in record length. A statistical figure of the dataset has been created to show more clarity on the number of audio records of similar length recordings as depicted in Figure 8.

5.2. Training

We first split the dataset into the following three subsets: 70% for training, 10% for validation, and 20% for testing. Before starting the training procedure to align the wave and text encoders, we pre-trained the VAE using only the audio waveforms. The VAE takes an input audio x and attempts to reconstruct x ^ after compressing x into z. To reduce the training time, memory usage, and complexity, we fed a randomly selected sliced hidden representation z with a window size of 32 to the wave decoder. To compute the STFT and LSGAN losses, we extracted the corresponding audio segments from the ground-truth raw waveforms as training targets. We followed the Parallel WaveGAN [36] for the details of the adversarial training.
After the VAE had converged, we utilized the pre-trained wave encoder to initiate the training procedure. The aim was to align the latent distribution of the text encoder with that of the pre-trained wave encoder so that it could later be recognized by the pre-trained wave decoder.
The training of both the VAE and the alignment of the wave and text encoders was conducted on two RTX A5000 GPUs manufactured by NVIDIA sourced from Denver, Colorado in United States of America. The VAE was trained using a batch size of 18 waveforms per GPU. The optimization was performed by utilizing the Adam optimizer [38] with the parameters set to β 1 = 0.9 , β 2 = 0.98 , and ϵ = 10 9 . The learning rate decay was scheduled by a factor of 0.9991 / 8 per epoch, starting from an initial value of 1 × 10 3 . It took 430 K steps for training until convergence.
The training for the alignment of the wave and text encoders was executed with a batch size of 12 sentences per GPU by utilizing the AdamW optimizer [39] with the parameters β 1 = 0.8 , β 2 = 0.98 , and a weight decay of λ = 0.01 . The learning rate decay followed a schedule of 0.9991 / 8 factors per epoch, starting from an initial learning rate of 2 × 10 4 . The training process reached convergence after 820 K steps.

6. Results and Discussion

In this section, we evaluate the performance of KTTS in terms of audio quality and inference speed.

6.1. Audio Quality

We evaluated the generated synthetic audio files in the test set to obtain the MOS to measure the audio quality. We kept the sentence content consistent among the different models so as to exclude other interference factors and avoid biases by only examining the audio quality. Each audio was listened to by at least 54 evaluators who were all native speakers of Central Kurdish. We compared the MOSs of the audio samples generated by our KTTS model with those of other well-known TTS models, which included (a) the GT (the ground-truth audio), (b) Tacotron 2 [1] (Mel + multi-band MelGAN [40]), (c) VITS [13], a conditional variational autoencoder with adversarial learning for end-to-end text-to-speech conversion, and (d) Glow TTS [20], a generative flow for text-to-speech conversion via monotonic alignment search (Mel + multi-band MelGAN). For each model, only ratings greater than one were considered, while those equal to or below this threshold were excluded from the analysis. The results are shown in Table 3. It can be seen that our KTTS outperformed the mentioned one-stage TTS system, and it reached the quality of the two-stage TTS systems.

6.2. Inference Speed

This section compares the inference speed of KTTS with that of both two-stage and one-stage systems. The comparison was conducted on a server with an “AMD Ryzen threadripper pro 3955wx” CPU with 16 cores, 256 GB of memory, and one NVIDIA RTX A5000 GPU with 24 GB of memory. Table 4 shows that the proposed model sped up the inference process by 8.32x compared with that of the one-stage VITS system [13]. Regarding the two-stage systems, the proposed model sped up the inference process by 47.49x with respect to Glow TTS [20] and by 53.73x with respect to Tacotron 2 [1], as the two-stage systems needed more processing time because two parallel models were included in their inference processes.
A real-time factor (RTF) comparison was conducted in order to evaluate the model’s efficiency in synthesizing speech. Table 5 shows that the proposed model outperformed the other one-stage and two-stage models in real time.
Since our approach relies on VAEs, it is important to acknowledge some inherent limitations of this method. VAEs, though promising in various applications, including TTS, face certain challenges. Being an unsupervised learning system, VAEs lack precise control over the speech features they generate. A key drawback of VAEs is the need to balance regularization and reconstruction accuracy, which can sometimes lead to distorted speech outputs. Additionally, the alignment between synthesized speech and input text is not explicitly defined in VAEs. Moreover, VAEs require meticulous tuning of several hyperparameters, such as the dimensionality of the latent space, the choice of a prior distribution, and the weighting of loss and reconstruction terms. These factors necessitate careful consideration to optimize performance and mitigate potential issues.
The proposed approach may yield suboptimal results under certain conditions, such as a lack of sufficient training data, complex phonetic variability, and real-time constraints. These conditions can involve complex computations and latent space sampling during inference, which may be computationally expensive.

7. Conclusions

This study introduced KTTS, an efficient end-to-end text-to-speech model tailored for generating high-quality Kurdish audio. By leveraging a pre-trained VAE for audio waveform reconstruction and integrating adversarial training techniques, we enhanced the expressiveness while ensuring high-fidelity speech synthesis. Our methodology effectively bridges the text-to-speech gap by aligning prior and posterior distributions within latent variables. An empirical evaluation on a custom dataset demonstrated KTTS’s superior performance, which was comparable to the ground truth and was validated through subjective human evaluation. This represents a significant advancement in TTS technology, offering efficiency, quality, and flexibility for the synthesis of Kurdish text and speech. Future work may focus on enhancing the model further and expanding its applicability to other Kurdish dialects, speech styles, and multi-speaker models. Another area for future work involves integrating the two training phases of the VAE and KTTS into a unified procedure, enabling seamless integration without the requirement to pre-train the VAE network’s encoder and decoder separately.

Author Contributions

H.A.A.: Data curation, original draft preparation, visualization, conceptualization, methodology, software, validation, resources, writing, reviewing and editing. T.A.R.: supervision, reviewing, and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data is accessible via the Mendeley Data repository at the following URL: https://data.mendeley.com/datasets/zhnvwsd7hs/1 (accessed on 18 June 2024).

Acknowledgments

We would like to acknowledge Jegr Nadhim for his invaluable assistance in preparing the server with two NVIDIA GPUs for our development process.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
TTSText to speech
MOSMean opinion score
VAEVariational autoencoder
GANsGenerative adversarial networks
KTTSKurdish text to speech

Appendix A

Appendix A.1. Kurdish Language

This section covers the three major Kurdish dialects and provides an extensive elaboration on the writing system of the Central Kurdish dialect. It also delves into the various pronunciation points and grammatical structure of the language. Most Kurds living in Iraq and Iran use the Central Kurdish dialect. It became an official language of Iraq in 2006. The Northern Kurdish dialect, on the other hand, is commonly spoken in areas such as Syria, Turkey, and Northwestern Iran, among others. Southern Kurdish is also known as Zazaki or Hawrami, and it can be heard in areas such as Ilam and Kermanshah in Iran. The main types of Kurdish are Central and Northern Kurdish, which are, respectively, written in the Latin and Arabic alphabets.
Despite the number of speakers, the number of resources for the Central Kurdish dialect is still more than sufficient. The system for writing the language was first established during the 1920s, and it has undergone numerous changes. Table A1 provides a brief overview of the phonological features of the phonetic alphabet system.
The writing system of Central Kurdish is similar to that of a phonemic system. The assigned letter of the language is referred to as a “phoneme”. Some exceptions are made. For example, the letter “Algorithms 17 00292 i001” is pronounced as palatal approximants “/j/”, and the “/i/” as it is referred to as a “vowel”. The letter “Algorithms 17 00292 i002” is also called a bilabial approximant, and it is pronounced as a a vowel: “/u/” or “/ʊ/”. The letter “Algorithms 17 00292 i002” can be written in a repetitive way, “Algorithms 17 00292 i003”, which is pronounced as “/u/”, a long vowel. Although it is one phoneme, it is not considered as a separate character in the keyboard layout released by the Department of Information Technology of the Regional Government of Kurdistan in 2014.
Although the Central Kurdish system is similar to that of Persian and Arabic writing, there are differences. Table A2 illustrates the differences between the three alphabets, as they can be used to distinguish among different languages. In Arabic and Persian, there are usually problems with the homograph and Kasre, but in Kurdish writing, these issues do not occur, unlike in other languages, such as Arabic, Persian, and English. Due to the mapping between the written and spoken terms, these issues are not as common.
Table A1. Central Kurdish phoneme list for the e-speak phonemizer.
Table A1. Central Kurdish phoneme list for the e-speak phonemizer.
Algorithms 17 00292 i004
Speech synthesis is a multidisciplinary field that has a wide range of problems. Among these is the issue with the management of unfamiliar words, as well as the prosody of proper nouns and foreign ones. In addition, synthesizers commonly encounter issues with the wave concatenation technique due to the varying effects of contextual and thematic factors [41].
It can be very challenging to create text due to the various language-related aspects involved. For instance, every non-standard term should have a phonetic equivalent. In addition, full words should be made from numbers and letters [42].
In Central Kurdish, some of the issues that can be encountered are hidden short-vowel characters. For instance, the words (دڵ، کردن، بردن) include hidden short vowels between consonant letters [43].
In contrast to other languages, such as Persian, English, and Arabic, the problems encountered by speech synthesizer users in Kurdish are not as common. Speech synthesizer users have to recognize the prosodic elements of written text, including the intonation, stress, and length. The features of continuous speech are influenced by the personality and emotions of the artist. Unfortunately, there is a lack of knowledge regarding the prosodic elements in written texts, which causes many of them to be frequently modified as speech is synthesized.
Table A2. Letters of Kurdish in comparison with Persian and Arabic letters [43].
Table A2. Letters of Kurdish in comparison with Persian and Arabic letters [43].
Algorithms 17 00292 i005

Appendix A.2. Phonemization

The phonemization process is the act of breaking words down into their phonemes. Phonemes are the smallest unit of speech sound, and they differentiate one word from another in a particular language. In our work, we used the e-speak phonemizer [34] to break the words down into their corresponding phonemes. We created a list of Central Kurdish letter phonemes that did not exist in the original repository. The phoneme list is illustrated in Table A1. As can be easily seen, we assigned some different phonemes from those in the IPA to comply with the e-speak package, as well as to overcome some pronunciation errors that we encountered during testing.

References

  1. Shen, J.; Pang, R.; Weiss, R.J.; Schuster, M.; Jaitly, N.; Yang, Z.; Chen, Z.; Zhang, Y.; Wang, Y.; Skerrv-Ryan, R.; et al. Natural tts synthesis by conditioning wavenet on mel spectrogram predictions. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 4779–4783. [Google Scholar]
  2. Oord, A.v.d.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K. Wavenet: A generative model for raw audio. arXiv 2016, arXiv:1609.03499. [Google Scholar]
  3. Kalchbrenner, N.; Elsen, E.; Simonyan, K.; Noury, S.; Casagrande, N.; Lockhart, E.; Stimberg, F.; Oord, A.; Dieleman, S.; Kavukcuoglu, K. Efficient neural audio synthesis. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 2410–2419. [Google Scholar]
  4. Ren, Y.; Ruan, Y.; Tan, X.; Qin, T.; Zhao, S.; Zhao, Z.; Liu, T.Y. Fastspeech: Fast, robust and controllable text to speech. arXiv 2019, arXiv:1905.09263. [Google Scholar]
  5. Peng, K.; Ping, W.; Song, Z.; Zhao, K. Non-autoregressive neural text-to-speech. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 7586–7598. [Google Scholar]
  6. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. arXiv 2014, arXiv:1406.2661. [Google Scholar]
  7. Li, N.; Liu, S.; Liu, Y.; Zhao, S.; Liu, M. Neural speech synthesis with transformer network. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 6706–6713. [Google Scholar]
  8. Kumar, K.; Kumar, R.; De Boissiere, T.; Gestin, L.; Teoh, W.Z.; Sotelo, J.; De Brebisson, A.; Bengio, Y.; Courville, A.C. Melgan: Generative adversarial networks for conditional waveform synthesis. arXiv 2019, arXiv:1910.06711. [Google Scholar]
  9. Kong, J.; Kim, J.; Bae, J. Hifi-gan: Generative adversarial networks for efficient and high fidelity speech synthesis. Adv. Neural Inf. Process. Syst. 2020, 33, 17022–17033. [Google Scholar]
  10. Ren, Y.; Hu, C.; Tan, X.; Qin, T.; Zhao, S.; Zhao, Z.; Liu, T.Y. Fastspeech 2: Fast and high-quality end-to-end text to speech. arXiv 2020, arXiv:2006.04558. [Google Scholar]
  11. Donahue, J.; Dieleman, S.; Bińkowski, M.; Elsen, E.; Simonyan, K. End-to-end adversarial text-to-speech. arXiv 2020, arXiv:2006.03575. [Google Scholar]
  12. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  13. Kim, J.; Kong, J.; Son, J. Conditional variational autoencoder with adversarial learning for end-to-end text-to-speech. arXiv 2021, arXiv:2106.06103. [Google Scholar]
  14. Ren, Y.; Tan, X.; Qin, T.; Zhao, Z.; Liu, T.Y. Revisiting oversmoothness in text to speech. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; Volume 1: Long Papers, pp. 8197–8213. [Google Scholar]
  15. Kong, J.; Park, J.; Kim, B.; Kim, J.; Kong, D.; Kim, S. VITS2: Improving Quality and Efficiency of Single-Stage Text-to-Speech with Adversarial Learning and Architecture Design. arXiv 2023, arXiv:2307.16430. [Google Scholar]
  16. Saiteja, K. Towards Building Controllable Text to Speech Systems. 2023. Available online: https://cdn.iiit.ac.in/cdn/cvit.iiit.ac.in/images/Thesis/MS/saiteja_kosgi/Sai_Thesis.pdf (accessed on 3 January 2024).
  17. Feng, X.; Yoshimoto, A. Llama-VITS: Enhancing TTS Synthesis with Semantic Awareness. arXiv 2024, arXiv:2404.06714. [Google Scholar]
  18. Casanova, E.; Shulby, C.; Gölge, E.; Müller, N.M.; De Oliveira, F.S.; Junior, A.C.; Soares, A.D.S.; Aluisio, S.M.; Ponti, M.A. SC-GlowTTS: An Efficient Zero-Shot Multi-Speaker Text-To-Speech Model. arXiv 2021, arXiv:2107.01423. [Google Scholar]
  19. Choi, B.J.; Jeong, M.; Lee, J.Y.; Kim, N.S. SNAC: Speakernormalized affine coupling layer in flow-based architecture for zero-shot multi-speaker text-to-speech. IEEE Signal Process. Lett. 2022, 29, 2502–2506. [Google Scholar] [CrossRef]
  20. Kim, J.; Kim, S.; Kong, J.; Yoon, S. Glow-tts: A generative flow for text-to-speech via monotonic alignment search. Adv. Neural Inf. Process. Syst. 2020, 33, 8067–8077. [Google Scholar]
  21. Łańcucki, A. Fastpitch: Parallel Text-to-Speech with Pitch Prediction. In Proceedings of the ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 6588–6592. [Google Scholar] [CrossRef]
  22. McAuliffe, M.; Socolof, M.; Mihuc, S.; Wagner, M.; Sonderegger, M. Montreal Forced Aligner: Trainable text-speech alignment using Kaldi. In Proceedings of the Interspeech, Stockholm, Sweden, 20–24 August 2017; pp. 498–502. [Google Scholar]
  23. Jeong, M.; Kim, H.; Cheon, S.J.; Choi, B.J.; Kim, N.S. Diff-tts: A denoising diffusion model for text-to-speech. arXiv 2021, arXiv:2104.01409. [Google Scholar]
  24. Chen, N.; Zhang, Y.; Zen, H.; Weiss, R.J.; Norouzi, M.; Dehak, N.; Chan, W. Wavegrad 2: Iterative refinement for text-to-speech synthesis. arXiv 2021, arXiv:2106.09660. [Google Scholar]
  25. Miao, C.; Zhu, Q.; Chen, M.; Ma, J.; Wang, S.; Xiao, J. EfficientTTS 2: Variational End-to-End Text-to-Speech Synthesis and Voice Conversion. IEEE ACM Trans. Audio Speech Lang. Process. 2024, 32, 1650–1661. [Google Scholar] [CrossRef]
  26. Li, Y.A.; Han, C.; Raghavan, V.; Mischler, G.; Mesgarani, N. StyleTTS 2: Towards Human-Level Text-to-Speech through Style Diffusion and Adversarial Training with Large Speech Language Models. arXiv 2023, arXiv:2306.07691. [Google Scholar]
  27. Bahrampour, A.; Barkhoda, W.; Azami, B.Z. Implementation of three text to speech systems for Kurdish language. In Iberoamerican Congress on Pattern Recognition; Springer: Heidelberg, Germany, 2009; pp. 321–328. [Google Scholar]
  28. Barkhoda, W.; ZahirAzami, B.; Bahrampour, A.; Shahryari, O.K. December. A comparison between allophone, syllable, and diphone based TTS systems for Kurdish language. In Proceedings of the 2009 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Louisville, KY, USA, 9–11 December 2009; pp. 557–562. [Google Scholar]
  29. Daneshfar, F.; Barkhoda, W.; Azami, B.Z. Implementation of a Text-to-Speech System for Kurdish Language. In Proceedings of the 2009 Fourth International Conference on Digital Telecommunications, Colmar, France, 20–25 July 2009; pp. 117–120. [Google Scholar]
  30. Hassani, H.; Kareem, R. Kurdish text to speech. In Proceedings of the KTTS Tenth International Workshop on Internationalisation of Products and Systems, Kuching, Malaysia, 11–14 July 2011; pp. 79–89. [Google Scholar]
  31. Fahmy, F.K.; Khalil, M.I.; Abbas, H.M. September. A transfer learning end-to-end arabic text-to-speech (tts) deep architecture. In Proceedings of the IAPR Workshop on Artificial Neural Networks in Pattern Recognition, Cham, Switzerland, 2–4 September 2020; pp. 266–277. [Google Scholar]
  32. Naderi, N.; Nasersharif, B.; Nikoofard, A. Persian speech synthesis using enhanced tacotron based on multi-resolution convolution layers and a convex optimization method. Multimed. Tools Appl. 2022, 81, 3629–3645. [Google Scholar] [CrossRef]
  33. Van Den Oord, A.; Vinyals, O. Neural discrete representation learning. arXiv 2017, arXiv:1711.00937. [Google Scholar]
  34. espeak-ng. 2024. Available online: https://github.com/espeak-ng/espeak-ng (accessed on 13 October 2023).
  35. Baevski, A.; Zhou, Y.; Mohamed, A.; Auli, M. wav2vec 2.0: A framework for self-supervised learning of speech representations. Adv. Neural Inf. Process. Syst. 2020, 33, 12449–12460. [Google Scholar]
  36. Yamamoto, R.; Song, E.; Kim, J.M. Parallel WaveGAN: A fast waveform generation model based on generative adversarial networks with multi-resolution spectrogram. In Proceedings of the ICASSP 2020–2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 6199–6203. [Google Scholar]
  37. Ahmad, H.A.; Rashid, T.A. Bridging the Gap: Central Kurdish Speech Corpus Construction and Recognition System Integration. Mendeley Data V1 2024. [Google Scholar] [CrossRef]
  38. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  39. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  40. Yang, G.; Yang, S.; Liu, K.; Fang, P.; Chen, W.; Xie, L. Multi-band MelGAN: Faster Waveform Generation for High-Quality Text-to-Speech. arXiv 2020, arXiv:2005.05106. [Google Scholar]
  41. Lemmetty, S. Review of Speech Synthesis Technology. Master’s Thesis, Helsinki University of Technology, Otaniemi, Finland, 1999. [Google Scholar]
  42. Macon, M.W. Speech Synthesis Based on Sinusoidal Modeling; Georgia Institute of Technology: Atlanta, GA, USA, 1996. [Google Scholar]
  43. Veisi, H.; Mohammad Amini, M.; Hosseini, H. Toward Kurdish language processing: Experiments in collecting and processing the AsoSoft text corpus. Digit. Scholarsh. Humanit. 2019, 35, 176–193. [Google Scholar] [CrossRef]
Figure 1. (A) Initially, a VAE is pre-trained using speech-to-speech data. During this phase, the VAE focuses on reconstructing the input speech waveform. (B) Training Procedure: This is the alignment phase where the pre-trained wave encoder of the VAE is utilized to ensure that the text encoder produces a distribution of latent variables identical to that generated by the wave encoder. (C) Inference Procedure: In this phase, the text encoder is trained to generate distributions that the wave decoder of the VAE can interpret and convert into speech waveforms.
Figure 1. (A) Initially, a VAE is pre-trained using speech-to-speech data. During this phase, the VAE focuses on reconstructing the input speech waveform. (B) Training Procedure: This is the alignment phase where the pre-trained wave encoder of the VAE is utilized to ensure that the text encoder produces a distribution of latent variables identical to that generated by the wave encoder. (C) Inference Procedure: In this phase, the text encoder is trained to generate distributions that the wave decoder of the VAE can interpret and convert into speech waveforms.
Algorithms 17 00292 g001
Figure 2. The text encoder featuring a modified transformer encoder with learnable positional encoding.
Figure 2. The text encoder featuring a modified transformer encoder with learnable positional encoding.
Algorithms 17 00292 g002
Figure 3. The positional encoder uses a 64-filter grouped 1D convolution to generate relative positional vectors.
Figure 3. The positional encoder uses a 64-filter grouped 1D convolution to generate relative positional vectors.
Algorithms 17 00292 g003
Figure 4. The wave encoder architecture utilizes a transformer structure akin to that of the text encoder, enhancing convergence during the application of KL divergence.
Figure 4. The wave encoder architecture utilizes a transformer structure akin to that of the text encoder, enhancing convergence during the application of KL divergence.
Algorithms 17 00292 g004
Figure 5. The feature encoder with a 2200-sample receptive field tokenizing 100 ms segments of raw waveforms for transformer processing.
Figure 5. The feature encoder with a 2200-sample receptive field tokenizing 100 ms segments of raw waveforms for transformer processing.
Algorithms 17 00292 g005
Figure 6. The wave decoder architecture inspired by WaveNet [2], featuring transposed 1D convolution, and dilated residual blocks, with adversarial training for enhanced waveform generation.
Figure 6. The wave decoder architecture inspired by WaveNet [2], featuring transposed 1D convolution, and dilated residual blocks, with adversarial training for enhanced waveform generation.
Algorithms 17 00292 g006
Figure 7. The architecture of the duration prediction model f d u r a t i o n .
Figure 7. The architecture of the duration prediction model f d u r a t i o n .
Algorithms 17 00292 g007
Figure 8. Histogram of Kurdish sentences in the dataset.
Figure 8. Histogram of Kurdish sentences in the dataset.
Algorithms 17 00292 g008
Table 1. Summary of the main points of the Kurdish literature review.
Table 1. Summary of the main points of the Kurdish literature review.
No.ReferenceYearMethodDatasetResult
1[27]2009Concatenative (Allophone, Syllable, and Diphone)Kurdish LanguageAllophone MOS 2.45
Syllable MOS 3.02
Diphone MOS 3.51
2[28]2009Concatenative (Allophone, Syllable, and Diphone)Kurdish LanguageBest quality score 3.5
Best DRT 97%
3[29]2009Concatenative (Allophone)Kurdish Language (2100 words)Best quality score 2.4
4[30]2011Concatenative (Diphone)Kurdish Language (2100 words)Best quality score 55%
5[31]2020Tacotron 2-Transfer LearningNawar Halabi’s Arabic Dataset(3 h) MOS 4.21
6[32]2022Tacotron 2Persian dataset(21 h) MOS 3.01–3.97
Table 2. Statistics of the sentences of the speech corpus.
Table 2. Statistics of the sentences of the speech corpus.
TopicsNo. of Sentences
News888
Sport631
Health463
Interview1240
Science65
Religion24
Economic275
General information224
Politics66
Education and literature1399
Article420
Social383
Total6078
Table 3. Comparison of the evaluated MOSs with 95% confidence intervals on the Gigant dataset.
Table 3. Comparison of the evaluated MOSs with 95% confidence intervals on the Gigant dataset.
ModelMOS (CI)
Ground Truth4.75 (±0.10)
Tacotron 2 (Mel + multi-band MelGAN)3.85 (±0.10)
Glow TTS (Mel + multi-band MelGAN)3.94 (±0.15)
VITS3.70 (±0.66)
KTTS3.94 (±0.16)
Table 4. A comparison of the inference speed with 95% confidence intervals.
Table 4. A comparison of the inference speed with 95% confidence intervals.
ModelLatency (s)Speedup (%)
One-Stage Systems
VITS0.560 ± 0.068/
KTTS0.517 ± 0.0938.32
Two-Stage Systems
Tacotron 2 (Mel + multi-band MelGAN)0.795 ± 0.011/
KTTS0.517 ± 0.09353.73
Glow TTS (Mel + multi-band MelGAN)0.763 ± 0.044/
KTTS0.517 ± 0.09347.49
Table 5. Comparison of RTF with 95% confidence intervals.
Table 5. Comparison of RTF with 95% confidence intervals.
ModelRTF (CI)
Tacotron 2 (Mel + multi-band MelGAN)0.099 ± 0.006
Glow TTS (Mel + multi-band MelGAN)0.095 ± 0.011
VITS0.0701 ± 0.006
KTTS0.065 ± 0.004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ahmad, H.A.; Rashid, T.A. Central Kurdish Text-to-Speech Synthesis with Novel End-to-End Transformer Training. Algorithms 2024, 17, 292. https://doi.org/10.3390/a17070292

AMA Style

Ahmad HA, Rashid TA. Central Kurdish Text-to-Speech Synthesis with Novel End-to-End Transformer Training. Algorithms. 2024; 17(7):292. https://doi.org/10.3390/a17070292

Chicago/Turabian Style

Ahmad, Hawraz A., and Tarik A. Rashid. 2024. "Central Kurdish Text-to-Speech Synthesis with Novel End-to-End Transformer Training" Algorithms 17, no. 7: 292. https://doi.org/10.3390/a17070292

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop