Figure - uploaded by Felix Gaschi
Content may be subject to copyright.
Comparison of domain and language specific models for translate-train in French.

Comparison of domain and language specific models for translate-train in French.

Source publication
Preprint
Full-text available
Natural language tasks like Named Entity Recognition (NER) in the clinical domain on non-English texts can be very time-consuming and expensive due to the lack of annotated data. Cross-lingual transfer (CLT) is a way to circumvent this issue thanks to the ability of multilingual large language models to be fine-tuned on a specific task in one langu...

Similar publications

Preprint
Full-text available
Open-source, multilingual medical large language models (LLMs) have the potential to serve linguistically diverse populations across different regions. Adapting generic LLMs for healthcare often requires continual pretraining, but this approach is computationally expensive and sometimes impractical. Instruction fine-tuning on a specific task may no...
Article
Full-text available
This paper presents a novel architecture for predicting the next word in bilingual Punjabi-English (BPE) social media texts. The goal is to enhance the performance and accuracy of next-word prediction in multilingual environments. Our proposed model, called NWP-CB (Next-Word Prediction using CNN and BiLSTM), incorporates two powerful deep learning...
Preprint
Full-text available
Bias studies on multilingual models confirm the presence of gender-related stereotypes in masked models processing languages with high NLP resources. We expand on this line of research by introducing Filipino CrowS-Pairs and Filipino WinoQueer: benchmarks that assess both sexist and anti-queer biases in pretrained language models (PLMs) handling te...
Article
Full-text available
In this work, we present a novel approach to lexical complexity prediction (LCP) that combines diverse linguistic features with encodings from deep neural networks. We explore the integration of 23 handcrafted linguistic features with embeddings from two well-known language models: BERT and XLM-RoBERTa. Our method concatenates these features before...