Hyungwon Yang’s scientific contributions

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (2)


Figure 3. The training speed of linear units per a single epoch
below describes how the encoder hyperparameters are initialized and the bold text indicates the default value.Table 1. The range of hyperparameters in the transformer encoder network
shows the types of hyperparameters used in the transformer network and their values. The default values of the hyperparameter in the box are bolded.
Hyperparameter experiments on end-to-end automatic speech recognition*
  • Article
  • Full-text available

March 2021

·

528 Reads

·

4 Citations

Phonetics and Speech Sciences

Hyungwon Yang

·

Hosung Nam
Download

Figure 1. The locations of the 6 primary articulatory organs, molar, and incisor. 
Figure 2. The structure of the speech inversion network 
Figure 4. The change of RMS in different numbers of layers. 'HN' means the numbers of hidden layers. 
Development of articulatory estimation model using deep neural network

September 2016

·

250 Reads

Phonetics and Speech Sciences

Heejo You

·

Hyungwon Yang

·

·

[...]

·

Hosung Nam

Speech inversion (acoustic-to-articulatory mapping) is not a trivial problem, despite the importance, due to the highly non-linear and non-unique nature. This study aimed to investigate the performance of Deep Neural Network (DNN) compared to that of traditional Artificial Neural Network (ANN) to address the problem. The Wisconsin X-ray Microbeam Database was employed and the acoustic signal and articulatory pellet information were the input and output in the models. Results showed that the performance of ANN deteriorated as the number of hidden layers increased. In contrast, DNN showed lower and more stable RMS even up to 10 deep hidden layers, suggesting that DNN is capable of learning acoustic-articulatory inversion mapping more efficiently than ANN.

Citations (1)


... The weights of CTC and Attention in the hybrid model were given by the hyper-parameter ctc weight . This parameter was left in its default value: γ CTC = ctc weight = 0.3 , because in 6 it was proved that this proportion is the best among other values. The weight of the attention mechanism is γ att = 0.7 according to (4). ...

Reference:

Multilingual end-to-end ASR for low-resource Turkic languages with common alphabets
Hyperparameter experiments on end-to-end automatic speech recognition*

Phonetics and Speech Sciences