A Nonlinearized Discriminant Analysis and its Application to Speech Impediment Therapy

András Kocsor, László Tóth and Dénes Paczolay

Abstract

This paper studies the application of automatic phoneme classification to the computer-aided training of the speech and hearing handicapped. In particular, we focus on how efficiently discriminant analysis can reduce the number of features and increase classification performance. A nonlinear counterpart of Linear Discriminant Analysis, which is a general purpose class specific feature extractor, is presented where the nonlinearization is carried out by employing the so-called 'kernel-idea'. Then, we examine how this nonlinear extraction technique affects the efficiency of learning algorithms such as Artificial Neural Network and Support Vector Machines.

1  Speech Impediment Therapy and Real-Time Phoneme Classification

This paper deals with the application of speech recognition to the computer-aided training of the speech and hearing handicapped. The program we present was designed to help in the speech training of the hearing impaired, where the goal is to support or replace their diminished auditory feedback with a visual one. But the program could also be applied to improving the reading skills of children with reading difficulties. Experience shows that computers more readily attract the attention of young people, who are usually more willing to practice with the computer than with the traditional drills.
Since both groups of our intended users consist mostly of young children it was most important that the design of the software interface be made attractive and novel. In addition, we realized early on that the real-time visual feedback the software provides must be kept simple, otherwise the human eye cannot follow it. Basically this is why the output of a speech recognizer seems better suited to this goal than the usual method where only the short-time spectrum is displayed: a few flickering discrete symbols are much easier to follow than a spectrum curve, which requires further mental processing. This is especially the case with very young children.
From the speech recognition point of view the need for a real-time output poses a number of special problems. Owing to the need for very fast classification we cannot delay the response even until the end of phonemes, hence we cannot employ complicated long-term models. The algorithm should process no more than a few neighbouring frames. Furthermore, since the program has to recognize vowels pronounced in isolation as well, a language model cannot be applied.
In our initial experiments we focussed on the classification of vowels, as the learning of the vowels is the most challenging for the hearing-impaired. The software supposes that the vowels are pronounced in isolation or in the form of two-syllable words, which is a more usual training strategy. The program provides a visual feedback on a frame-by-frame basis in the form of flickering letters, their brightness being proportional to the speech recognizer's output (see fig.1). To see the speaker's progress over longer periods, the program can also display the recognition scores during the previous utterance (see fig.2). Of course it is always possible to examine the sample spectra as well, either on a frame-by-frame or on an utterance-based basis. The utterances can be recorded and played back for further study and analysis by the teacher.
This article describes the experiments conducted with the LDA and Kernel-LDA transforms, intended to improve and possibly speed up the classification of vowels. As for the classification itself we used neural nets (ANN) and support vector machines (SVM). The section below explains the mathematical details of the Kernel-LDA transform, which is a new non-linear extension of the traditional LDA technique1.
Figure
Figure 1: A screenshot from EasySpeech. The real-time response of the system for vowel /a/.
Figure
Figure 2: A screenshot of EasySpeech after pronouncing the word /mimi/.

2  Linear Discriminant Analysis with and without Kernels

Before executing a learning algorithm it is a common practice to preprocess the data by extracting new features. Of the class specific feature extractors Linear Discriminant Analysis (LDA) is a traditional statistical method which has proved to be one of the most successful preprocessing techniques in classificaton2. The role of this method as preprocessing is twofold. Firstly it can improve classification performance, and secondly it may also reduce the dimensionality of the data and hence significantly speed up the classification.
The goal of Linear Discriminant Analysis is to find a new (not necessarily orthogonal) basis for the data which provides the optimal separation between groups of points (classes). Without loss of generality we will assume that the original data set, i.e. the input data lies in \mathdsRn, denoted by x1,¼,xr. The class label of each data vector is supposed to be known beforehand. Let us assume that we have k classes and an indicator function f():{1,¼,r}®{1,¼,k}, where f(i) gives the class label of the point xi. Let rj (j Î {1,¼,k}, r=r1+¼+rk) denote the number of vectors associated with label j in the data. In this section we now review the formalae for LDA, and also a nonlinear extension using the so-called 'Kernel-idea'.

2.1  Linear Discriminant Analysis

In order to extract m informative features from the n-dimensional input data, we first define a function t():\mathdsRn®\mathdsR which serves as a measure for selecting the m directions (i.e. base vectors of the new basis) one at a time. For a selected direction a a new real valued feature can be calculated as aT x. Intuitively, if larger values of t() indicate better directions and the chosen directions need to be somehow independent, choosing stationary points that have large values is a reasonable strategy. So we define a new basis for the input data based on m stationary points of t with dominant function values. Now let us define
t(a)= aT B a

aT W a
,       a Î \mathdsRn\{ 0 },
(1)
where B is the Between-class Scatter Matrix, while W is the Within-class Scatter Matrix. Here Between-class Scatter Matrix B represents the scatter of the class mean vectors mj around the overall mean vector m=[1/r]åi=1r xiT xi while the Within-class Scatter Matrix W shows the weighted average scatter of the covariance matrices Cj of the sample vectors having label j:
B
=
k
å
j=1 
rj

r
(mj-m)(mj-m)T
W
=
k
å
j=1 
rj

r
Cj
Cj
=
1

rj

å
f(i)=j 
(xi-mj)(xi-mj)T
      mj
=
1

rj

å
f(i)=j 
xi
(2)
Since t(a) is large when its nominator is large and its denominator is small, the within-class averages of the sample projected onto a are far from each other, while the variance of the classes is small. The larger the value of t(a) the farther the classes will be spaced and the smaller their spreads will be. It can be easily shown that stationary points of (1) correspond to the right eigenvectors of W-1B, where the eigenvalues form the corresponding function values. Since W-1B is not necessarily symmetrical the number of real eigenvalues can be less than n and the corresponding eigenvectors will not necessarily be orthogonal3 . If we select the m eigenvectors with the greatest real eigenvalues (denoted by a1,...,am), we will obtain new features from an arbitrary data vector y Î \mathdsRn by a1\negmedspaceT y,¼,am\negmedspaceT y .

2.2  Kernel-LDA

Here the symbol H denotes a real vector space that could be finite or infinite in dimension and we suppose a mapping F:\mathdsRn®H, which is not necessarily linear. In addition, let us assume that the algorithm of Linear Discriminant Analysis is denoted by P and its input is the points x1,¼,xr of the vector space \mathdsRn. The output of the algorithm is a linear transformation \mathdsRn®\mathdsRm, where both the degree of the dimension reduction (represented by m) and the n×m transformation matrix are determined by the algorithm itself. P(x1,¼,xr) will denote the transformation matrix which results from the input data. Then the algorithm P is replaced by an equivalent algorithm P¢ for which P(x1,¼,xr) = P¢(x1Tx1,¼,xiTxj,¼,xrTxr) holds for arbitrary x1,¼,xr. Thus P¢ is equivalent to P but its inputs are the pairwised dot products of the inputs of algorithm P. Then applying a nonlinear mapping F on the input data, yields a nonlinear feature transformation matrix P¢(F(x1)TF(x1),¼,F(xi)TF(xj),¼,F(xr)TF(xr)). These dot products can be computed here in H (which may be infinite in dimension), but if we have a low-complexity (perhaps linear) kernel function k():\mathdsRn×\mathdsRn®\mathdsR for which F(x)TF(y)=k(x,y),   x,y Î \mathdsRn, then Fxi)TF(xj) can also be computed with fewer operations (for example O(n)) even when the dimensions of F(xi) and F(xj) are infinite. So, after choosing a kernel function, the only thing that remains is to take the algorithm P¢ and replace the input elements x1Tx1, ¼, xiTxj, ¼, xrTxr with the elements k(x1,x1), ¼, k(xi,xj), ¼, k(xr,xr). The algorithm that arrises from this substitution can perform the transformations with a practically acceptable complexity, whatever the spatial dimension. This transformation (together with a properly chosen kernel function) results in a non-linear feature extraction. The key idea here is that we do not need to know the mapping F explicitly; we need only a kernel function k():\mathdsRn×\mathdsRn®\mathdsR for which there exists a mapping F such that F(x)TF(y)=k(x,y),   x,y Î \mathdsRn. There are many good publications about the proper choice of the kernel functions, and also about their theory in general[7]. The two most popular kernels are the following ( p Î \mathdsN+ and s Î \mathdsR+):
k1(x,y)=(xTy+1)p,    k2(x,y) = exp(-||x-y||2 / s).
(3)
Practically speaking, the original LDA algorithm is executed in a transformed (probably infinite) feature space H where the kernel function k gives implicit access to the elements of this space. In the following we present the kernel analogue of LDA by transforming the algorithm P to P¢. Let us consider the following function for a fixed k,F and H.
tF(a)= aT BF a

aT WF a
,        a Î H\{0},
(4)
where the matrices needed for LDA are now given in H:
BF
=
k
å
j=1 
rj

r
(mFj-mF)(mFj-mF)T
WF
=
k
å
j=1 
rj

r
CFj
CFj
=
1

rj

å
f(i)=j 
(F(xi)-mFj)(F(xi)-mFj)T
     mFj
=
1

rj

å
f(i)=j 
F(xi)
(5)
We may suppose without loss of generality that a=åi=1r ai [^(F)](xi) holds during the search for the stationary points of (4)4.
Now
aT BF a
=
æ
è
r
å
t=1 
at F(xt)T ö
ø
é
ë
k
å
j=1 
rj

r
æ
è
é
ë
1

rj

å
f(i)=j 
F(xi) ù
û
- é
ë
1

r
r
å
i=1 
F(xi) ù
û
ö
ø
æ
è
é
ë
1

rj

å
f(i)=j 
F(xi)T ù
û
- é
ë
1

r
r
å
i=1 
F(xi)T ù
û
ö
ø
ù
û
æ
è
r
å
s=1 
as F(xs) ö
ø
(6)
Since aT BF a can also be expressed as aT KBF a where a=[a1,¼,ar]T and where the matrix KBF is of size r×r, and for its elements KBFts with index (t,s) the following holds.:
KBFts
=
k
å
j=1 
rj

r
æ
è
é
ë
1

rj

å
f(i)=j 
k(xt,xi) ù
û
- é
ë
1

r
r
å
i=1 
k(xt,xi) ù
û
ö
ø
æ
è
é
ë
1

rj

å
f(i)=j 
k(xi,xs) ù
û
- é
ë
1

r
r
å
i=1 
k(xi,xs) ù
û
ö
ø
(7)
Then
aT WF a
=
æ
è
r
å
t=1 
at F(xt)T ö
ø
é
ë
k
å
j=1 
1

r

å
f(i)=j 
æ
è
F(xi) - é
ë
1

rj

å
f(i)=j 
F(xi) ù
û
ö
ø
æ
è
F(xi)T - é
ë
1

rj

å
f(i)=j 
F(xi)T ù
û
ö
ø
ù
û
æ
è
r
å
s=1 
as F(xs)T ö
ø
(8)
We can now express aT WF a in the form aT KWF a, where the matrix KWF is of size r×r and
KWFts
=
k
å
j=1 
1

r

å
f(i)=j 
æ
è
k(xt,xi) - é
ë
1

rj

å
f(i)=j 
k(xt,xi) ù
û
ö
ø
æ
è
k(xi,xs) - é
ë
1

rj

å
f(i)=j 
k(xi,xs) ù
û
ö
ø
(9)
Combining the above equations we obtain the equality
aT BF a

aT WF a
= aT KBF a

aTKWF a
(10)
This means that (4) can be expressed as dot products of F(x1),...,F(xr) and that the stationary points of this equation can be computed using the real eigenvectors5 of (KWF)-1KBF. We will use only those eigenvectors which correspond to the m dominant real eigenvalues, denoted by a1,¼,am. Consequently, the transformation matrix AF of Kernel-LDA is
é
ë
1

q1
r
å
i=1 
ai1 F(xi),¼, 1

qm
r
å
i=1 
aim F(xi) ù
û
,     qk= æ
è
r
å
i=1 
r
å
j=1 
aikajkk(xi,xj) ö
ø
1/2
 
,
(11)
where the value of the normalization parameter q is chosen such that the norm of the column vectors remains unity. For an arbitrary data vector y, new features can be computed via AFTF(y) =[[1/(q1)]åi=1r ai1k(xi,y),¼,[1/(qm)]åi=1r aimk(xi,y)].

3  Experimental Results

Corpus. For training and testing purposes we recorded samples from 25 speakers, mostly children aged between 8 and 15, but the database used contained some adults too. The speech signals were recorded and stored at a sampling rate of 22050 Hz in 16-bit quality. Each speaker uttered 59 two-syllable Hungarian words of the CVCVC form, where the consonants (C) were mostly unvoiced plosives to ease the detection of the vowels (V). The distribution of the vowels was approximately uniform in the database. Because we decided not to discriminate their long and short versions, we worked with 9 vowels althogether. In the experiments 20 speakers were used for training and 5 for testing.
Feature Sets. The signals were processed in 10 ms frames, the log-energies of 24 critical-bands being extracted using FFT and triangular weighting [5]. The energy of each frame was normalized separately, which means that only the spectral shape was used for classification. Our previous results showed that an additional cosine transform (which would lead to the most commonly used MFCC coefficients) does not affect the performance of the classifiers we had intended to apply, so it was omitted. Brief tests showed that neither varying the frame size nor increasing the number of filters gave any significant increase in classifier performance.
In our most basic tests we used only the filter-bank log-energies from the middle frame of the steady-state part of each vowel ("FBLE" set). Then we added the derivatives of these features to model the signal dynamics ("FBLE+Deriv" set). In another experiment we smoothed the feature trajectories to remove the effect of transient noises and disturbances ("FBLE Smooth" set). In yet another set of features we extended the log-energies with the gravity centers of four frequency bands, approximately corresponding to the possible values of the formants. These gravity centers allegedly give a crude approximation of the formants ("FBLA+Grav" set) [1]. Lastly, for the sake of curiosity we performed a test with the feature set of our segmental model("Segmental" set) [6]. This describes a whole phonemic segment rather than just one frame, it clearly could not be applied in a real-time system. So our aim then was simply to see the advantages of a segmental classifier over a frame-based one.
Classifiers. In all the experiments with Artificial Neural Nets (ANN) [2] the well-known three-layer feed-forward MLP networks were employed with the backpropagation learning rule. The number of hidden neurons was equal to the number of features.
In the Support Vector Machine (SVM) [7] experiments we always made use of the radial basis kernel function k2 (see eq. (3)).
Transformations. In our tests with LDA and Kernel-LDA the eigenvectors belonging to the 16 dominant eigenvalues were chosen as basis vectors for the transformed space and for Kernel-LDA the third-order polynomial kernel k1, where p=3 was used (see eq. (3)) .

4  Results and Discussion

Table 1 lists the recognition errors where the rows represent the five feature sets while the columns correspond to the applied transformation and classifier combinations.
On examining the results on the different feature sets we saw that adding the derivative did not increase performance. On the other hand smoothing the trajectories proved beneficial. Most likely a good combination of smoothing and derivation (or even better, RASTA filtering) would give better results.
As regards the gravity center features, they brought about on improvement, but only a slight one. This result accords with our previous experiments [3]. Lastly, the full segmental model clearly performed better than all the frame-based classifiers. This demonstrates the advantage of modeling full phonetic segments over frame-based classification.
When examining the effects of LDA and Kernel-LDA, it can be seen that a non-linear transformation normally performs better in separating the classes than its linear counterpart owing to its larger degree of freedom. One other interesting observation is that although the transformations retained only 16 features, the classifiers attain the same or better scores. Since the computation of LDA is fast, the reduction in the number of features speeds up not only the training but also the recognition phase. As yet, this does not hold for the Kernel-LDA algorithm we currently use, but we are working on a faster implementation.
Finally, as regards the classifiers, SVM consistently outperformed ANN by a few percentage. This can mostly be attributed to the fact that the SVM algorithm cope with overfitting, which is a common problem in ANN training.
none
ANN
none
SVM
LDA
ANN
(16)
LDA
SVM
(16)
K-LDA
ANN
(16)
K-LDA
SVM
(16)
FBLE (24)26.71 % 22.70 % 25.82 % 24.01 % 24.52 % 21.05 %
FBLE+Deriv (48) 25.82 % 24.01 % 27.30 % 24.34 % 24.34 % 21.21 %
FBLE+Grav (32) 24.01 % 22.03 % 24.67 % 23.85 % 22.87 % 20.72 %
FBLE Smooth (24) 23.68 % 21.05 % 23.03 % 21.87 % 22.70 % 19.90 %
Segmental (77) 19.57 % 19.08 % 20.04 % 18.42 % 18.09 % 17.26 %
Table 1: Recognition errors for the vowel classification task. The numbers in parenthesis correspond to the number of features.

5  Conclusion

Our results show that transforming the training data before learning can definitely increase classifier performance, and also speed up classification. We also saw that a non-linearized transformation is more effective than the traditional linear version, although they are currently much slower. At present we are working on a sparse data representation scheme that is hoped will give an order of magnitude increase in calculation speed. As regards the classifiers, SVM always performes slightly better than ANN, so we plan to employ it in the future. From the application point of view, our biggest problem at the moment is the feature set. We are looking for more phonetically-based features so as to decrease the classification error, since reliable performance is very important in speech impediment therapy.

References

[1]
Albesano, D., De Mori, R., Gemello, R., and Mana, F., A study on the effect of adding new dimensions to trajectories in the acoustic space, Proc. of EuroSpeech'99, pp. 1503-1506, 1999.
[2]
Bishop, C. M., Neural Networks for Pattern Recognition, Oxford University Press, 1995.
[3]
Kocsor, A., Tóth, L., Kuba, A. Jr., Kovács, K., Jelasity, M., Gyimóthy, T., and Csirik, J., A Comparative Study of Several Feature Transformation and Learning Methods for Phoneme Classification, Int. Journal of Speech Technology, Vol. 3., No. 3/4, pp. 263-276, 2000.
[4]
Mika, S., Rätsch, G., Weston, J., Schölkopf, B., and Müller, K.-R., Fisher Discriminant Analysis with Kernels, In Hu, Y.-H., Larsen, E., Wilson, E. and Douglas, S., editors, Neural Networks for Signal Processing IX, pages 41-48, IEEE, 1999.
[5]
Rabiner, L. R., Juang, B.-H., Fundamentals of Speech Recognition, Englewood Cliffs, NJ, Prentice Hall, 1993.
[6]
Toth, L., Kocsor, A., and Kovács, K., A Discriminative Segmental Speech Model and Its Application to Hungarian Number Recognition, In Sojka, P. et al.(eds.), Text, Speech and Dialogue, Proc. of TSD'2000, Springer Verlag LNAI Series, vol. 1902, pp. 307-313, 2000.
[7]
Vapnik, V. N., Statistical Learning Theory, John Wiley & Sons Inc., 1998.

Footnotes:

1 In [4] this method bears the name "Kernel Fisher Discriminant Analysis". Independently of these authors we arrived to the same formulae too, the only difference being that we derived the formulae for the multiclass case, naming the technique "Kernel-LDA". Although we recently reported our results of Kernel-LDA on word recognition in [6], the method itself was not described in great detail.
2One should note here that it can be directly used for classification as well.
3 Besides this, numerical problems can occure during the computation of W-1 if det(W) is near zero. The most probable cause for this could be the redundancy of feature components. But we know W is positive semidefinite. So if we add a small positive constant e to its diagonal, that is we work with W+eI instead of W, this matrix is guaranteed to be positive definite and hence should always be invertible. This small act of cheating can have only a negligible effect on the stationary points of (1).
4 This assumption can be arrived at in several ways, for instance we can decompose an arbitrary vector a into a1+a2, where a1 is the component of a which falls in SPAN([^(F)] (x1),¼,[^(F)] (xr)), while a2 gives the component perpendicular to it. Then from the derivation of t\negmedspace[^(F)] (a) it can be proved that a2\negmedspaceT\negmedspacea2=0 for stationary points.
5 Since in general KWF is a positive semidefinite matrix with its determinant sometimes near zero, it can be forced to be invertible using the technique presented in the subsection of LDA. Please see footnote 3 as well.


File translated from TEX by TTH, version 3.68.
On 20 Aug 2005, 16:02.