Shiyue Zhang’s research while affiliated with Tsinghua University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (15)


VV-Couplet: An open source Chinese couplet generation system
  • Conference Paper

November 2018

·

31 Reads

·

1 Citation

Jiyuan Zhang

·

Zheling Zhang

·

Shiyue Zhang

·


Collaborative Learning for Language and Speaker Recognition

February 2018

·

71 Reads

·

2 Citations

Communications in Computer and Information Science

Lantian Li

·

·

·

[...]

·

Shiyue Zhang

This paper presents a unified model to perform language and speaker recognition simultaneously and together. This model is based on a multi-task recurrent neural network, where the output of one task is fed in as the input of the other, leading to a collaborative learning framework that can improve both language and speaker recognition by sharing information between the tasks. The preliminary experiments presented in this paper demonstrate that the multi-task model outperforms similar task-specific models on both language and speaker tasks. The language recognition improvement is especially remarkable, which we believe is due to the speaker normalization effect caused by using the information from the speaker recognition component.




Medical Diagnosis From Laboratory Tests by Combining Generative and Discriminative Learning

November 2017

·

41 Reads

·

9 Citations

A primary goal of computational phenotype research is to conduct medical diagnosis. In hospital, physicians rely on massive clinical data to make diagnosis decisions, among which laboratory tests are one of the most important resources. However, the longitudinal and incomplete nature of laboratory test data casts a significant challenge on its interpretation and usage, which may result in harmful decisions by both human physicians and automatic diagnosis systems. In this work, we take advantage of deep generative models to deal with the complex laboratory tests. Specifically, we propose an end-to-end architecture that involves a deep generative variational recurrent neural networks (VRNN) to learn robust and generalizable features, and a discriminative neural network (NN) model to learn diagnosis decision making, and the two models are trained jointly. Our experiments are conducted on a dataset involving 46,252 patients, and the 50 most frequent tests are used to predict the 50 most common diagnoses. The results show that our model, VRNN+NN, significantly (p<0.001) outperforms other baseline models. Moreover, we demonstrate that the representations learned by the joint training are more informative than those learned by pure generative models. Finally, we find that our model offers a surprisingly good imputation for missing values.



Enhanced Neural Machine Translation by Learning from Draft

October 2017

·

46 Reads

Neural machine translation (NMT) has recently achieved impressive results. A potential problem of the existing NMT algorithm, however, is that the decoding is conducted from left to right, without considering the right context. This paper proposes an two-stage approach to solve the problem. In the first stage, a conventional attention-based NMT system is used to produce a draft translation, and in the second stage, a novel double-attention NMT system is used to refine the translation, by looking at the original input as well as the draft translation. This drafting-and-refinement can obtain the right-context information from the draft, hence producing more consistent translations. We evaluated this approach using two Chinese-English translation tasks, one with 44k pairs and 1M pairs respectively. The experiments showed that our approach achieved positive improvements over the conventional NMT system: the improvements are 2.4 and 0.9 BLEU points on the small-scale and large-scale tasks, respectively.


Memory-augmented Neural Machine Translation

August 2017

·

52 Reads

·

14 Citations

Neural machine translation (NMT) has achieved notable success in recent times, however it is also widely recognized that this approach has limitations with handling infrequent words and word pairs. This paper presents a novel memory-augmented NMT (M-NMT) architecture, which stores knowledge about how words (usually infrequently encountered ones) should be translated in a memory and then utilizes them to assist the neural model. We use this memory mechanism to combine the knowledge learned from a conventional statistical machine translation system and the rules learned by an NMT system, and also propose a solution for out-of-vocabulary (OOV) words based on this framework. Our experiments on two Chinese-English translation tasks demonstrated that the M-NMT architecture outperformed the NMT baseline by 9.0 and 2.7 BLEU points on the two tasks, respectively. Additionally, we found this architecture resulted in a much more effective OOV treatment compared to competitive methods.


Memory-augmented Chinese-Uyghur Neural Machine Translation

June 2017

·

102 Reads

·

5 Citations

Neural machine translation (NMT) has achieved notable performance recently. However, this approach has not been widely applied to the translation task between Chinese and Uyghur, partly due to the limited parallel data resource and the large proportion of rare words caused by the agglutinative nature of Uyghur. In this paper, we collect ~200,000 sentence pairs and show that with this middle-scale database, an attention-based NMT can perform very well on Chinese-Uyghur/Uyghur-Chinese translation. To tackle rare words, we propose a novel memory structure to assist the NMT inference. Our experiments demonstrated that the memory-augmented NMT (M-NMT) outperforms both the vanilla NMT and the phrase-based statistical machine translation (SMT). Interestingly, the memory structure provides an elegant way for dealing with words that are out of vocabulary.


Flexible and Creative Chinese Poetry Generation Using Neural Memory

May 2017

·

74 Reads

·

45 Citations

It has been shown that Chinese poems can be successfully generated by sequence-to-sequence neural models, particularly with the attention mechanism. A potential problem of this approach, however, is that neural models can only learn abstract rules, while poem generation is a highly creative process that involves not only rules but also innovations for which pure statistical models are not appropriate in principle. This work proposes a memory-augmented neural model for Chinese poem generation, where the neural model and the augmented memory work together to balance the requirements of linguistic accordance and aesthetic innovation, leading to innovative generations that are still rule-compliant. In addition, it is found that the memory mechanism provides interesting flexibility that can be used to generate poems with different styles.


Citations (12)


... The Multilingual Minorlingual Automatic Speech Recognition (M2ASR) project [1] aims to change the situation. The aims of this project are to construct a full set of speech and text resources for five minor languages (Tibetan, Mongolian, Uyghur, Kazakh, and Kirghiz) in the northwest of China, and make the resources open and free for research purposes. ...

Reference:

M2ASR-KIRGHIZ: A Free Kirghiz Speech Database and Accompanied Baselines
M2ASR: Ambitions and first year progress
  • Citing Conference Paper
  • November 2017

... The neural machine translation (NMT) models still struggle in the translation task on Uyghur-Chinese with complex morphology and limited resources [18]. To tackle this problem, Zhang et al. [19] proposed a novel memory structure to alleviate the rare word problem caused by the agglutinative nature of Uyghur. Pan et al. [20] proposed a multi-source neural model that employs two separate encoders to encode the Uyghur source word sequence and the Uyghur linguistic feature sequences. ...

Memory-augmented Chinese-Uyghur neural machine translation
  • Citing Conference Paper
  • December 2017

... Another way to give the decoder access to the full target-side context is the two-stage approach of Li et al. [172] who first drafted a translation, and then employed a multisource NMT system to generate the final translation from both the source sentence and the draft. Zhang et al. [173] proposed a similar scheme but generated the draft translations in reverse order. ...

Enhanced neural machine translation by learning from draft
  • Citing Conference Paper
  • December 2017

... When the duration of the test utterances is long (greater than 3 seconds), the i-vector approach achieves better performance, and if the test utterances is short (about 1 second) and the number of enrollment utterances is small (less than 10 utterances), the dvector approach performs better. Future work will investigate more powerful techniques such as x-vector [16], PTN [17] and LRE with auxiliary information [18], [19]. ...

Collaborative Learning for Language and Speaker Recognition
  • Citing Chapter
  • February 2018

Communications in Computer and Information Science

... Within the domain of FM-augmented NMT, various approaches have been implemented in the past, resulting in enhanced translation performance. Some examples include integrating FMs to the transformer-based NMT architectures through modifying the decoding process (Cao and Xiong, 2018;Gu et al., 2018;Khandelwal et al., 2021;Reheman et al., 2023), adding a lexical memory to the NMT architecture (Feng et al., 2017), attaching rewards for matched translation pieces from FMs into the NMT output layer (Zhang et al., 2018), introducing additional attention layers to capture relevant information from translation memories (TMs) (He et al., 2021), or modifying the whole architecture, enabling it to edit FMs to obtain a final translation (Gu et al., 2019;Bouthors et al., 2023). ...

Memory-augmented Neural Machine Translation
  • Citing Conference Paper
  • January 2017

... In the realm of medical diagnosis, generative models have demonstrated their superiority over traditional methods by providing accurate and reliable results [105,120,143]. These models excel in clinical text analysis, processing diagnostic records, medical documents, and literature to aid in disease understanding and diagnosis. ...

Medical Diagnosis From Laboratory Tests by Combining Generative and Discriminative Learning
  • Citing Article
  • November 2017

... One intuitive application is the open-domain QA, where it intrinsically necessitates retrieving relevant knowledge from outer sources since there is no supporting information at hand (Chen et al., 2017;Xu et al., 2021a,b). Another major application is neural machine translation with translation memory, where the memory can either be the bilingual training corpus (Feng et al., 2017;Gu et al., 2018) or a large collection of monolingual corpus (Cai et al., 2021). It also has received great attention in other text generation tasks including dialogue response generation (Cai et al., 2019; and knowledge-intensive generation , as well as some information extraction tasks including named entity recognition (Wang et al., 2021a), and relation extraction (Zhang et al., 2021). ...

Memory-augmented Neural Machine Translation
  • Citing Article
  • August 2017

... Each lyric was assessed by three individuals in a blind review manner, where the reviewers had no information about the generation method used for each lyric. Following previous work in generating poems [7,8] , we evaluated the generated lyrics based on three criteria: consistency, fluency, and meaning. Each criterion was rated from 1 to 3, representing bad, normal, and good, respectively. ...

Flexible and Creative Chinese Poetry Generation Using Neural Memory
  • Citing Conference Paper
  • January 2017

... Note that, as Cherokee and English have different word orders (English follows SVO; Cherokee has variable word orders), one Cherokee phrase could be translated into two English words that are far apart in the sentence. This increases the difficulty of SMT that relies on phrase correspondence and is not good at distant word reordering (Zhang et al., 2017). We implement our SMT systems by Moses (Koehn et al., 2007). ...

Memory-augmented Chinese-Uyghur Neural Machine Translation
  • Citing Article
  • June 2017

... The choice of the GRU is made to avoid the gradient vanishing problem of the LSTMs. GRU is more robust to noise [15] and outperforms the LSTM in several tasks [16]. Moreover, the GRUs are less computationally expensive as they have two gates unlike the LSTM which has four. ...

Memory visualization for gated recurrent neural networks in speech recognition
  • Citing Conference Paper
  • March 2017