Professional Documents
Culture Documents
An Articulatory-Based Singing Voice Synthesis Using
An Articulatory-Based Singing Voice Synthesis Using
An Articulatory-Based Singing Voice Synthesis Using
Figure 5: Example of ultrasound image and the same 4.2. LSF estimation
image corrupted by additional speckle noise of Selected variables were used as inputs of one MLP per
variance 0.04. LSF. We optimized the number of hidden units according to
the lowest validation error. The 12 MLP trained for the LSF
3.2. Autoencoder parameters, initializations and mapping had 26-33 input units, 1-7 hidden units and 1 output
training (each corresponding to one of the 12 LSF coefficients).
Because LSF coefficients are not in the range [0-1], we used a
A deep autoencoder, built by stacking RBMs, was used to sigmoid activation function between input and hidden layer
extract salient descriptors of articulatory data from lips and and a linear activation function between hidden layer and
tongue images. We used a symmetric autoencoder with 4 output.
hidden layers of size 200 for the separate RBMs and 100, 200
and 100 respectively for the shared RBMs. Autoencoder is 4.3. Comparison with state of the art methods
first randomly initialized before RBM training. In order to
improve the reconstruction performances of the autoencoder, In order to validate our nonlinear feature extraction method,
the maximum number of epochs during training was set to we compared its performances to the performances achieved
500. The size of mini-batches was set to 100. The choice of with a linear state-of-the art method based on the extraction of
these hyperparemeters was motivated by series of test showing the EigenLips and EigenTongues [4]. Inspired by EigenFaces
the relationship between parameters and quality of in [8], this method uses Principal Components Analysis
reconstructed LSF. An example of hyperparameter performed on tongue and lips images. The idea is to create a
optimization is given Table 1. The number of hidden units for finite set of orthogonal images, named EigenLips and
RBM 1 is chosen in order to minimize the mean Spectral EigenTongues. Every tongue or lips configuration can be
Distorsion (see section 5.1). represented in this space by projection over EigenTongues and
EigenLips. This representation encodes significant structural
Table 1 Optimization of the number of hidden units for informations about lips and tongue [4].
RBM 1 according to spectral distorsion scores.
5. Quality assessments
Number of hidden units for Mean Spectral Distorsion
RBM 1 (dB) 5.1. Quality measurements
100 4.6 Quality measurements were estimated on the 5000 examples
200 4.3 from the independent test set, without retraining. Output
400 4.5 predicted LSF were most of the time in the range of original
LSF. Direct comparisons between true LSF and their predicted
version using either EigenLips and EigneTongues or
Multimodal Autoencoder (see Figure 6 and Figure 7) is
4. LSF mapping completed by a numerical evaluation, the so-called spectral
distorsion .
4.1. Variable selection 0.4
LSF 1
Reference LSF
The mapping between hidden representation computed from Estimated LSF
prevents the algorithm from selecting features containing the Figure 6 Comparing true vs. estimated LSF in the test
same information, using the Gram-Schmidt orthogonalization set using Multimodal Autoencoder on the first LSF.
procedure. The aim of this algorithm is to select features
0.4
LSF 1
LSF were computed with audio signals, we can use LPC
Reference LSF
Estimated LSF
residuals as source reference. Noise activation will produce
whispered voice. EGG signals could not be used directly as a
0.35 source: this correlate of the glottal activity is too indirect and
noisy. We used some source information extracted from the
derivative of EGG signals (dEGG) using the Voicebox toolbox
0.3
[10] to estimate the parameters of a glottal flow model based
LSF value
〈 ∑ 10 log 〉 (1)