Content uploaded by Jordan J. Bird
Author content
All content in this area was uploaded by Jordan J. Bird on Mar 29, 2019
Content may be subject to copyright.
Accent Classification in Human Speech Biometrics
for Native and Non-native English Speakers
Jordan J. Bird
School of Engineering and Applied Science, Aston
University
Birmingham, UK
birdj1@aston.ac.uk
Elizabeth Wanner
School of Engineering and Applied Science, Aston
University
Birmingham, UK
e.wanner@aston.ac.uk
Anikó Ekárt
School of Engineering and Applied Science, Aston
University
Birmingham, UK
a.ekart@aston.ac.uk
Diego R. Faria
School of Engineering and Applied Science, Aston
University
Birmingham, UK
d.faria@aston.ac.uk
ABSTRACT
Accent classication provides a biometric path to high reso-
lution speech recognition. This preliminary study explores
various methods of human accent recognition through clas-
sication of locale. Classical, ensemble, timeseries and deep
learning techniques are all explored and compared. A set of
diphthong vowel sounds are recorded from participants from
the United Kingdom and Mexico, and then formed into a large
static dataset of statistical descriptions by way of their Mel-
frequency Cepstral Coecients (MFCC) at a sample window
length of 0.02 seconds. Using both at and timeseries data,
various machine learning models are trained and compared
to the scientic standard Hidden Markov Model (HMM). Re-
sults through 10 fold cross validation show that a vote of
average probabilities between a Random Forest and Long
Short-term Memory Neural Network result in a classication
accuracy of 94.74%, outperforming the speech classication
standard Hidden Markov Model by a 5% increase in accuracy.
CCS CONCEPTS
•Computing methodologies →Speech recognition
;
Ma-
chine learning approaches;Learning settings.
KEYWORDS
Computational Linguistics, Speech Recognition, Accent Recog-
nition, Machine Learning, Biometrics, Voice Assistants
1 INTRODUCTION
Speech recognition in the home is quickly becoming a more
viable and aordable technology through systems such as Ap-
ple Siri, Amazon Alexa and Google Home. Home assistants
perform many tasks such as purchasing products, remotely
controlling home appliances, and making phonecalls among
many countless other skills. Despite the growing abilities
and availability of Smart Homes and their respective devices,
there are several issues hampering their usage in terms of
the level of scientic state-of-the-art. Specically, non-native
English speakers often encounter issues when attempting
to converse with automated assistants [
1
,
12
,
31
], and thus
measures are required to be able to correctly recognise the
accent or locale of a speaker, which can then be logically
acted on accordingly.
In this work, a dataset of spoken sounds from the English
phonetic dictionary are grouped based on the locale of the
speaker. Speakers are both native (West Midlands, UK; Lon-
don, UK) and non-native (Mexico City, MX; Chihuahua, MX)
English speakers producing a four-class problem. Various
single, ensemble and deep learning models are trained and
compared in terms of their classication ability for accent
recognition. A at dataset of 26 200ms Mel-frequency Cep-
stral Coecients form data objects for classication, except
for a timeseries of the aforementioned datapoints that are
generated for Hidden Markov Model training and prediction.
The main contributions of this work are as follows:
•
A benchmark of the most common model used for
contemporary voice recognition, the Hidden Markov
Model, when training from a uniform spoken audio
dataset and producing predictions of speaker accent/locale.
•
Single and ensemble Models are presented for the clas-
sication of accent of two Mexican locales and two
British locales.
•
The nal comparison of the eleven machine learning
models in which a vote of average probabilities of Ran-
dom Forest and LSTM is suggested as the best model
with a very high classication accuracy of 94.74%.
2 BACKGROUND
Related Works
Hidden Markov Models (HMM), since their inception in 1966,
remain a modern approach for speech recognition due to
their retaining of eectiveness given more computational
resources. An earlier work from 1996 found that, using 5
hidden Markov states due to the computational resources
available at the time, four spoken accents could be classied
at an observed accuracy of 74.5%[
3
]. It must be noted that far
deeper exploration into optimal HMM topology and struc-
ture is now possible due to the larger degree of processing
power available to researchers in the modern day. A more
modern work found that Support Vector Machines (SVM)
and HMM could classify between three dierent national
locales (Chinese, Indian, Canadian) at an average accuracy
of 93.8% for both models [
30
], though, this study only classi-
ed data from male speakers due to the statistical frequency
dierences between gender and voice.
Long Short Term Memory neural networks are often suc-
cesfully experimented with in terms of accent recognition.
A related experiment found accuracies at around 50% when
classifying 12 dierent locales [
20
]. Of the dataset gathered
from multiple locations across the globe, it was observed
that the highest recall rates were that of Japanese, Chinese,
and German scoring 65%, 61% and 60% respectively. Sub-
jects were recorded in their native language. An alternative
network-based approach, Convolutional Neural Networks,
were used to classify speech from English, Korean and Chi-
nese dialects at an average accuracy of 88%[
28
]. A proposed
approach to the accent problem in speech recognition, also
using a CNN, oered a preliminary study into deriving a
conversion matrix to be applied to Mel-frequency Cepstral
Coecients which would act to translate the user’s accent
into a specied second accent before speech recognition is
performed[5].
The eectiveness of voting between multiple models to
create a simple classier ensemble has been observed to
be extremely eective in many human-machine interaction
domains such as Sentiment Analysis [
9
] and EEG brainwave
classication [7].
It is worth noting in terms of criticism, that many accent
recognition experiments rarely dene the spoken language
itself, often resulting in the classication of a subject speak-
ing their native language in their natural locale. Through
this, it is very possible that the classiers in question would
not only learn from accent, but from natural language pat-
terns as a form of audible Natural Language Processing, since
such eects are also represented within MFCC data. For the
goal of improving voice recognition for non-native English
speakers who are speaking English, the previous models
would be somewhat worthless or inaccurate. The originality
of this experiment is to classify data retrieved from native
and non-native English speakers (who are all requested to
pronounce sounds from the English phonetic dictionary, as
if they were speaking in English), with the ultimate goal of
providing a path to improving voice recognition on English
language devices and services for non-native speakers.
Mel-frequency Cepstral Coeicients
Soundwaves are complex, random, and non-stationary and
thus classication of raw sound is very dicult. For example,
real-time monitoring of sound data would simply give a
measured frequency at a single point in time, which would be
impossible to classify since the behaviour of the wave is not
described. A sliding time windowing technique is introduced
and a statistical extraction is performed based on the section
of the wave appearing within the observed time window.
This results in a set of temporal mathematical descriptions of
wave sections. Mel-frequency Cepstral Coecient (MFCC)
of the sound is often cited as the most eective statistical
modelling method of sound waves[
23
,
27
]. MFCC datasets
are produced from a sliding time window as follows:
(1)
The Fourier Transform (FT) of the time window data
is derived:
X(ω)=
+∞
∫
−∞
x(t)e−jωtdt.(1)
(2)
The powers from the FT are mapped to the mel-scale,
which is the psychological scale of audible pitch[29].
(3)
The Mel-frequency Cepstrum (MFC), or power spec-
trum of sound, is considered and logs of each of their
powers are taken.
(4)
The derived mel-log powers are treated as a signal,
and a discrete cosine transform (DCT) is measured:
Xk=
N−1
Õ
n=0
cos π
N(n+1
2)kk=0, . . ., N−1.(2)
The MFCCs are nally considered as the resultant amplitudes
of the spectrum generated through the above process. A
mathematical description of a short wave section has been
generated, and provide attributes for the mapped class.
3 METHOD
A dataset was gathered for concurrent experiments and
made freely available online for future experiments
1
, from
the accent locales observed in Fig. 1. The voice recognition
dataset contained seven individual phonetic sounds spoken
ten times each by subjects from the United Kingdom and
Mexico. Those from the UK were native English speakers
whereas those from Mexico were native Spanish and uent
English speakers who were asked to pronounce the phonetic
sounds as if they were speaking English. 26 logs of MFCC
data were extracted from each dataset at a sliding time win-
dow of 200ms, each data object were the 26 MFCC features
1
https://www.kaggle.com/birdy654/speech-recognition-dataset-england-
and-mexico
2
Figure 1: Accent Locale of Experimental Subjects (Not to
Scale). To the left, Chihuahua (topmost) and Mexico City
(boom-most) are located, in Mexico; to the right the West
Midlands (topmost) and London (boom-most) are located,
in the United Kingdom.
2 4 6 8 10 12 14 16 18 20 22 24 26
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
MFCC Log
Information Gain
Figure 2: Information Gain of Each MFCC Log Attribute in
the Dataset
mapped to the accent of the speaker. Accents were sourced
from the West Midlands and London in the UK whereas
accents from Mexico were sourced from Mexico City and
Chihuahua. Weights of all four classes were balanced (since
the clips diered in length) to simulate an equally distributed
dataset. The dataset was formatted into a timeseries (rela-
tional attributes) for HMM training and prediction. The In-
formation Gain classication ability of each of the individual
attributes are shown in in Fig. 2. Models were all trained
on an AMD FX8320 8-core processor with a clock speed of
3.5GHz.
An extremely complex dense neural network was trained
for purposes of results and comparison, but would be unreal-
istically complex for actual real world use. A neural network
of two hidden layers (256, 128) was trained on the CUDA
cores of an NVidia GTX680 for 1000 epochs per fold, with
a batch size of 100. This model was trained using the Keras
library [
13
] running via the TensorFlow Platform. This is
labelled as"DENSE NN" in the results.
Due to the large degree of computational resources re-
quired for training several of the selected models, 10-fold
cross validation was chosen for model averaging.
Chosen Machine Learning Methods
Various methods of machine learning were selected based on
their ranging statistical[
32
] and methodological dierences,
as to aptly benchmark a selection of methods for spoken
accent classication. This section briey details the scientic
method followed by each of the methods to derive knowledge
through learning for classication.
Random Trees and Forests. A Decision Tree is a data structure
of conditional control statements based on attribute values,
which are then mapped to a tree. Classication is performed
by cascading a data point down the tree through each con-
ditional check until a leaf node (a node with no remaining
branches), which is mapped to a Class, ie. the prediction of
the model. The growth of the tree is based on the entropy of
its end node, that is, the level of disorder in classes found on
that node. Entropy of a node is considered as:
E(S)=−
c
Õ
i=1
Pi×loд2Pi,(3)
where the entropies of each class prediction are measured at
a leaf node.
A Random Tree (RT) is a method of generating a ran-
dom decision tree generated in which k-random attributes
are selected at each control statement as well as a best-t
decision[
25
]. An overtted tree is generated for the input set
and therefore cross-validation or a test-set are required for
proper measurement of prediction ability. J48 is an algorithm
to generate a decision tree based on C4.5[
26
]. Rather than
randomness, information entropy is used to calculate a best
split at each node ie. the most optimal split at that given step.
C4.5 requires far more processing power than RT due to the
requirement of this calculation.
A Random Forest is an ensemble of many Random Trees
through Bootstrap Aggregation (bagging) and Voting[
19
].
Numerous RT’s are generated by a random selection of data.
During classication, all of the trees will vote on their pre-
diction, and the majority vote is selected as the overall pre-
diction. Random Forests tend to outperform Random Trees
due to their decreasing of variance without increasing of the
model bias.
Bayesian Classifiers. Bayes Theorem[
4
] is the comparable
probability that data point dwill match to Class C. The
3
Figure 3: Diagram of a Long Short-term Memory Network
Unit [18]
theorem is given as follows:
P(A|B)=P(B|A)P(A)
P(B),(4)
where the probability of P(A) being true is related to the
probability of the H with evidence P(A|B). In terms of this
work, this would take inputs of MFCC measurements and
attempt to classify the spoken accent via selecting that of
the highest probability based on previous evidence.
Naive Bayes classication is given as follows:
ˆ
y=k∈(1,. . ., K)p(Ck)
n
Ö
i=1
p(xi|Ck),(5)
where class label yis given to data object k. The naivety in
Bayesian algorithms concerns the assumed independence of
attribute values (or existence), whether or not the assumption
holds true for a data.
Long Short-term Memory. Long Short Term Memory (LSTM)
is a form of Articial Neural Network in which multiple
Recurrent Neural Networks (RNN) will predict based on
state and previous states. As seen in Fig. 3, the data structure
of a neuron within a layer is an ’LSTM Block’. The general
idea is as follows:
The rst step requires the LSTM block to select stored
data to delete:
ft=σ(Wf.[ht=1,xt+bf).(6)
Where Wf are the weights of the blocks, h is the previous
output of the block (t-1), xt are the inputs received by the
block, and a bias is applied via bf.
The block must then select which data to store/remember.
Based on cell input i,Ct are the values generated.
ot=σ(Wi.[ht=1,xt+bi),(7)
˜
Ct=tanh(Wc.[ht=1,xt+bc).(8)
The block will then update parameters through a convolu-
tional operation:
Ct=ft∗Ct−1+it∗˜
Ct.(9)
Finally, output Ot is produced, and the hidden state is
updated:
ot=σ(Wo.[ht=1,xt+bo),(10)
ht=ot∗tanh(Ct).(11)
Due to the observed consideration of time sequences, i.e.
previously seen data, it is often found that time dependent
data are very eectively classied due to the memory state
of the block. LSTM ANN is thus a particularly powerful
technique in terms of speech recognition [
17
] and brainwave
classication [
7
,
11
,
15
] - since both are temporal, wave-like
data.
Nearest Neighbour Classification. K-nearest Neighbour (KNN)
is a method of classication based on measured distance from
ktraining data points [
2
]. KNN is considered a lazy learning
technique since all computation is deferred and only required
during the classication stage. KNN is performed as follows:
(1)
Convert nominal attributes to integers mapped to the
attribute label
(2) Normalise attributes
(3)
Map all training data to n-dimensional space where n
are are the values of attributes
(4) Lazy Computation Stage - For each data point:
(a)
Plot the data point to the previously generated n-
dimensional space
(b)
Have K-nearest points all vote on the point based on
their value
(c)
Predict the class of the point with that which has
received the highest number of votes
Logistic Regression. Logistic Regression is a process of sym-
metric statistics where a numerical value is linked to a prob-
ability of event occurring, ie. the number of driving lessons
to predict pass or fail [33].
In a two class problem within a dataset containing inum-
ber of attributes and
β
model parameters, the log odds lis
derived via
l=β0+Íx
i=0βi+xi
and the odds of an outcome
are shown through
o=bβ0+Íx
i=0βi+xi
which can be used to
predict an outcome based on previous observation.
Support Vector Machines. Support Vector Machines (SVM)
classify data points by optimising a data-dimensional hyper-
plane to most aptly separate them, and then classifying based
on the distance vector measured from the hyper-plane[
14
].
Optimisation follows the goal of the average margins be-
tween points and the separator to be at the maximum pos-
sible value. Generation of an SVM is performed through
Sequential Minimal Optimisation (SMO), a high-performing
algorithm to generate and implement an SVM classier[
24
].
4
To perform this, the large optimisation problem is broken
down into smaller sub-problems, these can then be solved
linearly. For multipliers a, reduced constraints are given as:
0≤a1,a2≤C,
y1,a1+y2,a2=k,
(12)
where there are data classes yand kis the negative of the
sum over the remaining terms of the equality constraint.
Hidden Markov Models. Markov Chains are a probabilis-
tic model that describe a sequence and probability of an
event occurring based on those which have been observed
previously[
16
]. Each previously observed event is repre-
sented as a Hidden Unit and therefore the most optimal
number of hidden states required is largely data dependent.
The general idea of the HMM process is as follows:
Y=y(0),y(1), ...y(L−1),(13)
denotes the probability of event Yoccurring based on the
the sequence of length L. Secondly,
P(Y)=Õ
X
P(Y|X)P(X),(14)
describes the probability of Y where the sum runs over all of
the generated hidden node sequences, given as X:
X=x(0),x(1), ..., x(L−1).(15)
Classication is nally chosen based on highest probabil-
ity on previously studied data sequences within the hidden
model through the Bayesian process [4].
Voting. Voting is a simple method of fusing the decisions
of multiple classiers and calculating an output prediction.
For example, an ensemble of two classiers with diering
classication abilities may possible produce a better result
when working together, ie. selected for their strengths. Vot-
ing is performed by various metrics, in which all classiers
will vote for a class, and then a prediction is produced by the
highest vote. Methods of voting include:
•
Vote weighted by probability to classify an individual
class
•Give 1 vote to predicted Class - Majority
•Vote based on overall ability to classify
•Vote weighted by condence
•Vote based on Min and Max Probabilities
The class is predicted based purely on the maximum vote
received.
4 RESULTS
Manual Tuning
Hidden Markov Units as well as hidden LSTM units were
linearly explored. Preliminary experimentation found that a
single layer of LSTM units persistently outperformed deeper
25 50 75 100 125 150 175 200
86
86.5
87
87.5
88
88.5
89
89.5
90
HMM Hidden Units
Classication Accuracy (%)
Figure 4: Exploration of HMM Hidden Unit Selection
25 50 75 100
90
92
94
91.51
92 92.01 92.01
LSTM Hidden Units
Classication Accuracy (%)
Figure 5: Exploration of LSTM Hidden Unit Selection
networks, and thus only one layer was linearly searched. The
chosen amount of HMM hidden units was selected as 200
since it had the superior classication accuracy of 89.65% as
observed in Figs 4 and 5 respectively. The chosen amount of
hidden units for the LSTM were selected as 75 since it too
had the most superior classication accuracy of 92.01%.
Overall Results
Table 1 displays the overall classication accuracy of the
selected single models when predicting the locale of the
speaker at each 200ms audio interval. The best single model
was an LSTM with 92.01% accuracy, closely followed by the
extremely complex dense neural network for benchmark
5
Table 1: Single Classier Results for Accent Classication (Sorted Lowest to Highest)
Model NB BN J48 LR RT SVM HMM RF KNN(10) DENSE NN LSTM
Accuracy 58.29 70.62 85.2 85.8 85.94 86.19 89.65 89.72 90.76 91.55 92.01
Table 2: Democratic Voting Processes for Ensemble Classi-
cation
Model Accuracy (%)
Democracy RF, LSTM KNN, LSTM KNN, RF
Avg. Prob. 94.74 94.63 92.62
Product Prob. 94.73 94.62 92.62
purposes, and then the K-Nearest Neighbours, and Hidden
Markov Models.
The best ensemble, and overall best, was a vote of average
probabilities between the Random Forest and LSTM, achiev-
ing 94.74% accuracy, this can be seen in the exploration of
democratic voting processes with best models, in Table 2.
5 CONCLUSION AND NEXT STEPS
This study has shown the eectiveness of various machine
learning techniques in terms of classifying the accent of
the subject based on recorded audio data. The diphthong
phoneme sounds were succesfully classied into four dif-
ferent accents from the UK and Mexico with an accuracy
of 94.74% when a manually tuned LSTM of 200 units and
a Random Forest are ensembled through a vote of average
probabilities.
Leave-one-out (LOO) cross validation has been observed
to be superior to test-set and k-fold cross-validation tech-
niques but requires far more processing time[
21
], this study
would have been around 3000 times more complex due to
there being 30,000 classiable data objects. It is likely better
results would be attained through this approach but with the
resources available, was not possible. Furthermore, more in-
tense searching of the problem spaces of HMM and LSTM hid-
den unit selection should be performed since relatively large
dierences were observed in minute topological changes.
Evolutionary algorithms have been observed to be a strong
method of topology selection and tuning [6, 8, 10, 22].
ACKNOWLEDGMENTS
REFERENCES
[1]
Neal Alewine, Eric Janke, Paul Sharp, and Roberto Sicconi. 2008. Sys-
tems and methods for building a native language phoneme lexicon
having native pronunciations of non-native words derived from non-
native pronunciations. US Patent 7,472,061.
[2]
Naomi S Altman. 1992. An introduction to kernel and nearest-neighbor
nonparametric regression. The American Statistician 46, 3 (1992), 175–
185.
[3]
Levent M Arslan and John HL Hansen. 1996. Language accent clas-
sication in American English. Speech Communication 18, 4 (1996),
353–367.
[4]
Thomas Bayes, Richard Price, and John Canton. 1763. An essay towards
solving a problem in the doctrine of chances. (1763).
[5]
Amy Bearman, Kelsey Josund, and Gawan Fiore. [n. d.]. Accent Con-
version Using Articial Neural Networks. ([n. d.]).
[6]
Jordan J. Bird, , Elizabeth Wanner, Aniko Ekart, and Diego R. Faria.
2019. Phoneme Aware Speech Recognition through Evolutionary
Optimisation. In The Genetic and Evolutionary Computation Conference.
GECCO.
[7]
Jordan J. Bird, Aniko Ekart, Christopher D. Buckingham, and Diego R.
Faria. 2019. Mental Emotional Sentiment Classication with an EEG-
based Brain-Machine Interface. In The International Conference on
Digital Image and Signal Processing (DISP’19). Springer.
[8]
Jordan J. Bird, Aniko Ekart, and Diego R. Faria. 2019. Evolutionary
Optimisation of Fully Connected Articial Neural Network Topology.
In SAI Computing Conference 2019. SAI.
[9]
Jordan J. Bird, Aniko Ekart, and Diego R. Faria. 2019. High Resolution
Sentiment Analysis by Ensemble Classication. In SAI Computing
Conference 2019. SAI.
[10]
Jordan J. Bird, Diego R. Faria, Luis J. Manso, Aniko Ekart, and Christo-
pher D. Buckingham. 2019. A Deep Evolutionary Approach to Bioin-
spired Classier Optimisation for Brain-Machine Interaction. Com-
plexity 2019 (2019). https://doi.org/10.1155/2019/4316548
[11]
Jordan J. Bird, Luis J. Manso, Eduardo P. Ribiero, Aniko Ekart, and
Diego R. Faria. 2018. A Study on Mental State Classication using
EEG-based Brain-Machine Interface. In 9th International Conference
on Intelligent Systems. IEEE.
[12]
William Byrne, Eva Knodt, Sanjeev Khudanpur, and Jared Bernstein.
1998. Is automatic speech recognition ready for non-native speech?
A data collection eort and initial experiments in modeling conversa-
tional Hispanic English. Proc. Speech Technology in Language Learning
(STiLL) 1, 99 (1998), 8.
[13] François Chollet et al. 2015. Keras. https://keras.io.
[14]
Corinna Cortes and Vladimir Vapnik. 1995. Support-vector networks.
Machine learning 20, 3 (1995), 273–297.
[15]
PR Davidson, RD Jones, and MTR Peiris. 2006. Detecting behavioral
microsleeps using EEG and LSTM recurrent neural networks. In 2005
IEEE Engineering in Medicine and Biology 27th Annual Conference. IEEE,
5754–5757.
[16]
Paul A Gagniuc. 2017. Markov Chains: From Theory to Implementation
and Experimentation. John Wiley & Sons.
[17]
Alex Graves, Navdeep Jaitly, and Abdel-rahman Mohamed. 2013. Hy-
brid speech recognition with deep bidirectional LSTM. In Automatic
Speech Recognition and Understanding (ASRU), 2013 IEEE Workshop on.
IEEE, 273–278.
[18]
Klaus Gre, Rupesh K Srivastava, Jan Koutník, Bas R Steunebrink,
and Jürgen Schmidhuber. 2017. LSTM: A search space odyssey. IEEE
transactions on neural networks and learning systems 28, 10 (2017),
2222–2232.
[19]
Tin Kam Ho. 1995. Random decision forests. In Document analysis and
recognition, 1995., proceedings of the third international conference on,
Vol. 1. IEEE, 278–282.
6
[20]
Yishan Jiao, Ming Tu, Visar Berisha, and Julie M Liss. 2016. Accent Iden-
tication by Combining Deep Neural Networks and Recurrent Neural
Networks Trained on Long and Short Term Features.. In Interspeech.
2388–2392.
[21]
Ron Kohavi et al
.
1995. A study of cross-validation and bootstrap for
accuracy estimation and model selection. In Ijcai, Vol. 14. Montreal,
Canada, 1137–1145.
[22]
Alejandro Martín, Raúl Lara-Cabrera, Félix Fuentes-Hurtado, Valery
Naranjo, and David Camacho. 2018. EvoDeep: A new evolutionary
approach for automatic Deep Neural Networks parametrisation. J.
Parallel and Distrib. Comput. 117 (2018), 180–191.
[23]
Lindasalwa Muda, Mumtaj Begam, and Irraivan Elamvazuthi. 2010.
Voice recognition algorithms using mel frequency cepstral coecient
(MFCC) and dynamic time warping (DTW) techniques. arXiv preprint
arXiv:1003.4083 (2010).
[24]
John Platt. 1998. Sequential minimal optimization: A fast algorithm
for training support vector machines. (1998).
[25]
Anantha M Prasad, Louis R Iverson, and Andy Liaw. 2006. Newer
classication and regression tree techniques: bagging and random
forests for ecological prediction. Ecosystems 9, 2 (2006), 181–199.
[26] J Ross Quinlan. 2014. C4. 5: programs for machine learning. Elsevier.
[27]
Md Sahidullah and Goutam Saha. 2012. Design, analysis and experi-
mental evaluation of block based transformation in MFCC computation
for speaker recognition. Speech Communication 54, 4 (2012), 543–565.
[28] Corey Shih. [n. d.]. Speech Accent Classication. ([n. d.]).
[29]
Stanley Smith Stevens, John Volkmann, and Edwin B Newman. 1937.
A scale for the measurement of the psychological magnitude pitch.
The Journal of the Acoustical Society of America 8, 3 (1937), 185–190.
[30]
Hong Tang and Ali A Ghorbani. 2003. Accent classication using
support vector machine and hidden markov model. In Conference of
the Canadian Society for Computational Studies of Intelligence. Springer,
629–631.
[31]
Laura Mayeld Tomokiyo and Alex Waibel. 2003. Adaptation methods
for non-native speech. Multilingual Speech and Language Processing 6
(2003).
[32]
Jessica PM Vital, Diego R Faria, Gonçalo Dias, Micael S Couceiro,
Fernanda Coutinho, and Nuno MF Ferreira. 2017. Combining discrimi-
native spatiotemporal features for daily life activity recognition using
wearable motion sensing suit. Pattern Analysis and Applications 20, 4
(2017), 1179–1194.
[33]
Strother H Walker and David B Duncan. 1967. Estimation of the
probability of an event as a function of several independent variables.
Biometrika 54, 1-2 (1967), 167–179.
7