Douglas W. Oard’s research while affiliated with University of Maryland Global Campus and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (376)


An Interdisciplinary Approach to Human-Centered Machine Translation
  • Preprint
  • File available

June 2025

·

33 Reads

Marine Carpuat

·

·

·

[...]

·

François Yvon

Machine Translation (MT) tools are widely used today, often in contexts where professional translators are not present. Despite progress in MT technology, a gap persists between system development and real-world usage, particularly for non-expert users who may struggle to assess translation reliability. This paper advocates for a human-centered approach to MT, emphasizing the alignment of system design with diverse communicative goals and contexts of use. We survey the literature in Translation Studies and Human-Computer Interaction to recontextualize MT evaluation and design to address the diverse real-world scenarios in which MT is used today.

Download

Fig. 1. Sampling distribution of each domain in each session and simulated stream. The color of the bars represents the session, along with the session ID marked at the bottom of each bar.
Fig. 2. Success@5 of each model updating strategy where the x-axis is the running session ID. Values in parenthesis in the legend are macro-average of Success@5 across all query sets and streams. Differences between each strategy are all statistically significant using the paired t-test described in Section 4.4 except for the pair of CF w/o Replay and MURR-LM after multiple test corrections.
Fig. 3. Query set breakdown of five model updating strategies in Scenario D2. Each subgraph is the effectiveness over sessions of the set of queries introduced in the session indicated in the title. Values in parenthesis are macro-average Success@5 on query sets in only D2. Values at the x-axis is the running session ID.
Fig. 4. Ablation study on the number of replay triples and the regularization strength under MURR-CF in Scenario D2. The x-axis is the session ID. The dashed gray line is the setup (200 triples, α = 0.01) used in the main experiments. Using 0 replay triples is essentially CF w/o Replay in the main results.
Dataset Statistics of LoTTe Forum Datasets.
MURR: Model Updating with Regularized Replay for Searching a Document Stream

April 2025

·

7 Reads

The Internet produces a continuous stream of new documents and user-generated queries. These naturally change over time based on events in the world and the evolution of language. Neural retrieval models that were trained once on a fixed set of query-document pairs will quickly start misrepresenting newly-created content and queries, leading to less effective retrieval. Traditional statistical sparse retrieval can update collection statistics to reflect these changes in the use of language in documents and queries. In contrast, continued fine-tuning of the language model underlying neural retrieval approaches such as DPR and ColBERT creates incompatibility with previously-encoded documents. Re-encoding and re-indexing all previously-processed documents can be costly. In this work, we explore updating a neural dual encoder retrieval model without reprocessing past documents in the stream. We propose MURR, a model updating strategy with regularized replay, to ensure the model can still faithfully search existing documents without reprocessing, while continuing to update the model for the latest topics. In our simulated streaming environments, we show that fine-tuning models using MURR leads to more effective and more consistent retrieval results than other strategies as the stream of documents and queries progresses.









Translate-Distill: Learning Cross-Language Dense Retrieval by Translation and Distillation

March 2024

·

14 Reads

·

7 Citations

Lecture Notes in Computer Science

Prior work on English monolingual retrieval has shown that a cross-encoder trained using a large number of relevance judgments for query-document pairs can be used as a teacher to train more efficient, but similarly effective, dual-encoder student models. Applying a similar knowledge distillation approach to training an efficient dual-encoder model for Cross-Language Information Retrieval (CLIR), where queries and documents are in different languages, is challenging due to the lack of a sufficiently large training collection when the query and document languages differ. The state of the art for CLIR thus relies on translating queries, documents, or both from the large English MS MARCO training set, an approach called Translate-Train. This paper proposes an alternative, Translate-Distill, in which knowledge distillation from either a monolingual cross-encoder or a CLIR cross-encoder is used to train a dual-encoder CLIR student model. This richer design space enables the teacher model to perform inference in an optimized setting, while training the student model directly for CLIR. Trained models and artifacts are publicly available on Huggingface.


Citations (64)


... However, since documents and queries can be generated by different types of agents, the original generative assumption, and its conceptual generalization just proposed, do not hold. This state-ofaffairs is reminiscent of that in cross-lingual retrieval where queries and documents are written in different languages [26,43]. Accordingly, re-visiting the generative theory to relevance is an interesting future direction to explore in multi-agent retrieval settings. ...

Reference:

A Multi-Agent Perspective on Modern Information Retrieval
Cross-language Retrieval
  • Citing Chapter
  • December 2024

... The recent emergence of large language models (LLMs) has enabled automated, reliable nuggetbased evaluation (Pradeep et al., 2024;Alaofi et al., 2024;Pradeep et al., 2025b;Abbasiantaeb et al., 2025). Several RAG evaluation frameworks-such as FactScore (Min et al., 2023), RUBRIC (Farzi and Dietz, 2024), and others Mayfield et al., 2024)-incorporate the nugget concept, although most of these proposed approaches are either not validated or primarily validated on traditional ad hoc retrieval, and hence their applicability to longform answers is unclear. We refer readers to Pradeep et al. (2025b) for a more detailed discussion of related work. ...

On the Evaluation of Machine-Generated Reports
  • Citing Conference Paper
  • July 2024

... In contrast, there is little work on neural dual encoders in such a streaming scenario. Recent work by Lawrie et al. [21] investigated incrementally indexing a large corpus using PLAID-X [40] with a fixed model, but only evaluated on a fixed set of queries at the end of the stream. How the model performs during and across the stream remains unknown. ...

PLAID SHIRTTT for Large-Scale Streaming Dense Retrieval
  • Citing Conference Paper
  • July 2024

... A parallel direction to our approach is recent literature leveraging distillation techniques. Yang et al. [40,39] recently proposed Translate-Distill, where a strong cross-encoder is leveraged as a teacher to distil its knowledge more efficient biencoder models in both a CLIR and MLIR setting as an alternative to translatetrain. Zhuang et al. [45] propose to augment the documents' representations with queries generated using mT5 in languages other than the original document language to mitigate the data scarcity of CLIR. ...

Translate-Distill: Learning Cross-Language Dense Retrieval by Translation and Distillation
  • Citing Chapter
  • March 2024

Lecture Notes in Computer Science

... Recent advancements focus on end-to-end retrieval systems that eliminate translation dependencies [19,22], improving scalability in multilingual settings [17]. Techniques such as language model pretraining [39], data curation and translation [22,23], and knowledge distillation [19,37,38] have enhanced the effectiveness of neural retrieval models [16,37]. ...

BLADE: Combining Vocabulary Pruning and Intermediate Pretraining for Scaleable Neural CLIR
  • Citing Conference Paper
  • July 2023

... In recent years, there has also been a growing interest in evaluating cross-lingual information retrieval. HC4 [15] and HC3 [16] were recently proposed to benchmark IR methods when dealing with Chinese, Persian, and Russian documents. Similarly, the TREC NeuCLIR track [13,14] released both NeuCLIR 1, a collection of Persian, Chinese and Russian Common Crawl News documents as well as neuMARCO, a cross-language machine-translated version of MSMARCO to study the effect of neural approaches to CLIR. ...

HC3: A Suite of Test Collections for CLIR Evaluation over Informal Text
  • Citing Conference Paper
  • July 2023

... HC4 [15] and HC3 [16] were recently proposed to benchmark IR methods when dealing with Chinese, Persian, and Russian documents. Similarly, the TREC NeuCLIR track [13,14] released both NeuCLIR 1, a collection of Persian, Chinese and Russian Common Crawl News documents as well as neuMARCO, a cross-language machine-translated version of MSMARCO to study the effect of neural approaches to CLIR. Both of these collections are now standard benchmarks for evaluating the cross-language capabilities of IR methods. ...

Overview of the TREC 2022 NeuCLIR Track

... Within academic institutions, language technologies can facilitate seamless knowledge retrieval from diverse sources and languages, mitigating cultural and technical barriers (Litschko et al., 2022). Consequently, AI-based programming for information retrieval is essential in advancing comprehension and development within multilingual environments (Lawrie et al., 2023). Furthermore, AI-driven information retrieval systems significantly contribute to the efficiency and accuracy of knowledge dissemination. ...

Neural Approaches to Multilingual Information Retrieval
  • Citing Chapter
  • March 2023

Lecture Notes in Computer Science

... However, those who focus on the adaptation of pre-existing schemes, proposed, in essence, modalities of application of the principle of cultural hospitality. As Choi et al. (2022) point out, cultural hospitality constitutes "an approach to improve information systems by providing ethical resource descriptions and access" (554). In this way, "cultural hospitality refers to the ability of a system to connect existing knowledge with perspectives, expectations, and assumptions from different cultures and users" (554). ...

Cross‐cultural Information Access
  • Citing Article
  • October 2022

Proceedings of the Association for Information Science and Technology

... Although financial incentives can help with recruitment, they can also increase the likelihood of fraud (Bowen et al., 2008;Rogers et al., 2022). If the study design includes a financial reimbursement for participation, consider methods for compensating people that would make it more difficult or impossible for fraudsters or bots. ...

A User Study in a Pandemic: Some Lessons Learned
  • Citing Article
  • October 2022

Proceedings of the Association for Information Science and Technology