Jitao Xu’s scientific contributions

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (13)


Figure 2: BiSync editor displaying paraphrases (top) and translation alternatives for a given prefix (bottom).
Figure 3: Default BiSync settings.
BiSync: A Bilingual Editor for Synchronized Monolingual Texts
  • Preprint
  • File available

June 2023

·

24 Reads

·

Jitao Xu

·

François Yvon

In our globalized world, a growing number of situations arise where people are required to communicate in one or several foreign languages. In the case of written communication, users with a good command of a foreign language may find assistance from computer-aided translation (CAT) technologies. These technologies often allow users to access external resources, such as dictionaries, terminologies or bilingual concordancers, thereby interrupting and considerably hindering the writing process. In addition, CAT systems assume that the source sentence is fixed and also restrict the possible changes on the target side. In order to make the writing process smoother, we present BiSync, a bilingual writing assistant that allows users to freely compose text in two languages, while maintaining the two monolingual texts synchronized. We also include additional functionalities, such as the display of alternative prefix translations and paraphrases, which are intended to facilitate the authoring of texts. We detail the model architecture used for synchronization and evaluate the resulting tool, showing that high accuracy can be attained with limited computational resources. The interface and models are publicly available at https://github.com/jmcrego/BiSync and a demonstration video can be watched on YouTube at https://youtu.be/_l-ugDHfNgU .

Download



BLEU scores for the multi-domain test sets broken down by the edit distance ∆ betweeñ e and e. Each column represents a range of distances. N denotes the number of sentences in each group.
TER scores on multi-domain test sets. All is computed by concatenating test sets from all domains, with 11k sentences in total. Copy refers to copying the similar translation in the output. +R implies using the related segments instead of a full initial sentence for inference. Best performance in each block are in bold.
Bilingual Synchronization: Restoring Translational Relationships with Editing Operations

October 2022

·

15 Reads

Machine Translation (MT) is usually viewed as a one-shot process that generates the target language equivalent of some source text from scratch. We consider here a more general setting which assumes an initial target sequence, that must be transformed into a valid translation of the source, thereby restoring parallelism between source and target. For this bilingual synchronization task, we consider several architectures (both autoregressive and non-autoregressive) and training regimes, and experiment with multiple practical settings such as simulated interactive MT, translating with Translation Memory (TM) and TM cleaning. Our results suggest that one single generic edit-based system, once fine-tuned, can compare with, or even outperform, dedicated systems specifically trained for these tasks.


BLEU and COMET scores for each domain, the task is standard MT with sim > 0.6. All is computed by concatenating all test sets (11k sentences in total). Copy refers to copying the TM match into the output.
BLEU and COMET scores for each domain, the task is MT with TMs with sim > 0.6. All is computed by concatenating all test sets (11k sentences in total). Copy refers to copying the TM match into the output.
Non-Autoregressive Machine Translation with Translation Memories

October 2022

·

73 Reads

Non-autoregressive machine translation (NAT) has recently made great progress. However, most works to date have focused on standard translation tasks, even though some edit-based NAT models, such as the Levenshtein Transformer (LevT), seem well suited to translate with a Translation Memory (TM). This is the scenario considered here. We first analyze the vanilla LevT model and explain why it does not do well in this setting. We then propose a new variant, TM-LevT, and show how to effectively train this model. By modifying the data presentation and introducing an extra deletion operation, we obtain performance that are on par with an autoregressive approach, while reducing the decoding load. We also show that incorporating TMs during training dispenses to use knowledge distillation, a well-known trick used to mitigate the multimodality issue.


Figure 1: A graphical view of various captioning and subtitling strategies. T refers to transcripts. C and S respectively denote captions and subtitles.
Joint Generation of Captions and Subtitles with Dual Decoding

May 2022

·

35 Reads

Jitao Xu

·

François Buet

·

·

[...]

·

François Yvon

As the amount of audio-visual content increases, the need to develop automatic captioning and subtitling solutions to match the expectations of a growing international audience appears as the only viable way to boost throughput and lower the related post-production costs. Automatic captioning and subtitling often need to be tightly intertwined to achieve an appropriate level of consistency and synchronization with each other and with the video signal. In this work, we assess a dual decoding scheme to achieve a strong coupling between these two tasks and show how adequacy and consistency are increased, with virtually no additional cost in terms of model size and training complexity.




Figure 1: A graphical view of the dual decoder.
One Source, Two Targets: Challenges and Rewards of Dual Decoding

September 2021

·

55 Reads

Machine translation is generally understood as generating one target text from an input source document. In this paper, we consider a stronger requirement: to jointly generate two texts so that each output side effectively depends on the other. As we discuss, such a device serves several practical purposes, from multi-target machine translation to the generation of controlled variations of the target text. We present an analysis of possible implementations of dual decoding, and experiment with four applications. Viewing the problem from multiple angles allows us to better highlight the challenges of dual decoding and to also thoroughly analyze the benefits of generating matched, rather than independent, translations.


Can You Traducir This? Machine Translation for Code-Switched Input

May 2021

·

21 Reads

Code-Switching (CSW) is a common phenomenon that occurs in multilingual geographic or social contexts, which raises challenging problems for natural language processing tools. We focus here on Machine Translation (MT) of CSW texts, where we aim to simultaneously disentangle and translate the two mixed languages. Due to the lack of actual translated CSW data, we generate artificial training data from regular parallel texts. Experiments show this training strategy yields MT systems that surpass multilingual systems for code-switched texts. These results are confirmed in an alternative task aimed at providing contextual translations for a L2 writing assistant.


Citations (5)


... In the context of specialised NMT models, previous studies showed that FM-augmented NMT models attain their maximum potential in high-resource, domain-specific scenarios characterised by the availability of large bilingual datasets, which enhance the likelihood of retrieving FMs with higher similarity levels (Bulté and Tezcan, 2019;Tezcan and Bulté, 2022;Xu et al., 2023;Reheman et al., 2023). To address this limitation, some efforts have been undertaken to leverage additional monolingual data in the target language for directly retrieving similar translations through employing multilingual sentence embeddings, resulting in further improvements in translation quality (Cai et al., 2021;Tamura et al., 2023). ...

Reference:

Improving Fuzzy Match Augmented Neural Machine Translation in Specialised Domains through Synthetic Data
Integrating Translation Memories into Non-Autoregressive Machine Translation
  • Citing Conference Paper
  • January 2023

... However, different from it, we center more on neural machine translation and inspect this field with decent insight from two new perspectives: dependency management on the target side and training arrangement for NAT models. Other than delivering quantitative description and qualitative comparison for various methods, we also anticipate promising future directions for this area, following up the latest findings, including simultaneous translation [15,16], automatic speech recognition [17][18][19], speech translation [20][21][22], image caption [23], and text editing [24][25][26][27][28][29], as well as the emerging large language models [30,31]. ...

Bilingual Synchronization: Restoring Translational Relationships with Editing Operations
  • Citing Conference Paper
  • January 2022

... After selecting LEV as our best timestamp projection method, we evaluate cascade and direct ST systems trained in the same data condition. Before this, to ensure the competitiveness of our cascade baseline, we compare it with the results obtained on the MuST-Cinema test set by the other cascade systems presented in literature, namely: en→{de, fr} by , and en→fr by Xu et al. (2022). As these works report only BLEU with breaks, i.e., BLEU computed including also <eob> and <eol>, we compare our cascade baseline with them on that metric. ...

Joint Generation of Captions and Subtitles with Dual Decoding
  • Citing Conference Paper
  • January 2022

... • Mitigating the data scarcity problem of multi-parallel data by finetuning the dual decoder model from standard translation models; • A new parameter sharing scheme, where the two decoders share all their parameters and achieve comparable performance at a reduced model size; • Concrete solutions to mitigate the exposure bias problem between two decoders; • Quantitative evaluations of the increased consistency incurred by a tight interaction between decoders. The contributions in this chapter are published in (Xu and Yvon, 2021b;Xu et al., 2022a). ...

One Source, Two Targets: Challenges and Rewards of Dual Decoding
  • Citing Conference Paper
  • January 2021

... Synthetic datasets have also introduced codeswitching mainly based on words. These include random replacements based on words (Rijhwani et al., 2017;Xu and Yvon, 2021;Rizvi et al., 2021;Tarunesh et al., 2021) and replacements based on connected components of aligned words (Iyer et al., 2023). However, word-based switching may not completely reflect the code-switching phenomenon. ...

Can You Traducir This? Machine Translation for Code-Switched Input
  • Citing Conference Paper
  • January 2021