Alexander Fraser’s research while affiliated with Cyberknife Center Munich and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (140)


Figure 1: EMOBENCH-UA is a benchmark of basic emotions-Joy, Anger, Fear, Disgust, Surprise, Sadness, or None-detection in Ukrainian texts.
Figure 4: EMOBENCH-UA statistics per sets and emotions.
Figure 6: Confusion matrices of the top performing models fine-tuned on the EMOBENCH-UA training data.
EmoBench-UA: A Benchmark Dataset for Emotion Detection in Ukrainian
  • Preprint
  • File available

May 2025

·

1 Read

·

·

Alexander Fraser

While Ukrainian NLP has seen progress in many texts processing tasks, emotion classification remains an underexplored area with no publicly available benchmark to date. In this work, we introduce EmoBench-UA, the first annotated dataset for emotion detection in Ukrainian texts. Our annotation schema is adapted from the previous English-centric works on emotion detection (Mohammad et al., 2018; Mohammad, 2022) guidelines. The dataset was created through crowdsourcing using the Toloka.ai platform ensuring high-quality of the annotation process. Then, we evaluate a range of approaches on the collected dataset, starting from linguistic-based baselines, synthetic data translated from English, to large language models (LLMs). Our findings highlight the challenges of emotion classification in non-mainstream languages like Ukrainian and emphasize the need for further development of Ukrainian-specific models and training resources.

Download

Figure 4: Global risks ranked by severity over the short and long term. Reproduction of Figure FIGURE C from: the Global Risks Report 2025 .
Figure 5: Guidelines for the authors of the paper. Please reach out for any clarification.
NLP for Social Good: A Survey of Challenges, Opportunities, and Responsible Deployment

May 2025

·

19 Reads

Antonia Karamolegkou

·

·

Eunjung Cho

·

[...]

·

Recent advancements in large language models (LLMs) have unlocked unprecedented possibilities across a range of applications. However, as a community, we believe that the field of Natural Language Processing (NLP) has a growing need to approach deployment with greater intentionality and responsibility. In alignment with the broader vision of AI for Social Good (Toma\v{s}ev et al., 2020), this paper examines the role of NLP in addressing pressing societal challenges. Through a cross-disciplinary analysis of social goals and emerging risks, we highlight promising research directions and outline challenges that must be addressed to ensure responsible and equitable progress in NLP4SG research.


EXECUTE: A Multilingual Benchmark for LLM Token Understanding

May 2025

·

3 Reads

The CUTE benchmark showed that LLMs struggle with character understanding in English. We extend it to more languages with diverse scripts and writing systems, introducing EXECUTE. Our simplified framework allows easy expansion to any language. Tests across multiple LLMs reveal that challenges in other languages are not always on the character level as in English. Some languages show word-level processing issues, some show no issues at all. We also examine sub-character tasks in Chinese, Japanese, and Korean to assess LLMs' understanding of character components.


From Unaligned to Aligned: Scaling Multilingual LLMs with Multi-Way Parallel Corpora

May 2025

·

1 Read

Continued pretraining and instruction tuning on large-scale multilingual data have proven to be effective in scaling large language models (LLMs) to low-resource languages. However, the unaligned nature of such data limits its ability to effectively capture cross-lingual semantics. In contrast, multi-way parallel data, where identical content is aligned across multiple languages, provides stronger cross-lingual consistency and offers greater potential for improving multilingual performance. In this paper, we introduce a large-scale, high-quality multi-way parallel corpus, TED2025, based on TED Talks. The corpus spans 113 languages, with up to 50 languages aligned in parallel, ensuring extensive multilingual coverage. Using this dataset, we investigate best practices for leveraging multi-way parallel data to enhance LLMs, including strategies for continued pretraining, instruction tuning, and the analysis of key influencing factors. Experiments on six multilingual benchmarks show that models trained on multiway parallel data consistently outperform those trained on unaligned multilingual data.


Data-Efficient Hate Speech Detection via Cross-Lingual Nearest Neighbor Retrieval with Limited Labeled Data

May 2025

·

1 Read

Considering the importance of detecting hateful language, labeled hate speech data is expensive and time-consuming to collect, particularly for low-resource languages. Prior work has demonstrated the effectiveness of cross-lingual transfer learning and data augmentation in improving performance on tasks with limited labeled data. To develop an efficient and scalable cross-lingual transfer learning approach, we leverage nearest-neighbor retrieval to augment minimal labeled data in the target language, thereby enhancing detection performance. Specifically, we assume access to a small set of labeled training instances in the target language and use these to retrieve the most relevant labeled examples from a large multilingual hate speech detection pool. We evaluate our approach on eight languages and demonstrate that it consistently outperforms models trained solely on the target language data. Furthermore, in most cases, our method surpasses the current state-of-the-art. Notably, our approach is highly data-efficient, retrieving as small as 200 instances in some cases while maintaining superior performance. Moreover, it is scalable, as the retrieval pool can be easily expanded, and the method can be readily adapted to new languages and tasks. We also apply maximum marginal relevance to mitigate redundancy and filter out highly similar retrieved instances, resulting in improvements in some languages.


Figure 1: Performance of zero-/few-shot prompted LLMs vs. fine-tuned XLM-T across varying training sizes.
Figure 2: Performance of zero-/few-shot prompted LLMs vs. fine-tuned XLM-T across varying training sizes.
Complete Zero-and Few-shot Prompting Results for Spanish.
Can Prompting LLMs Unlock Hate Speech Detection across Languages? A Zero-shot and Few-shot Study

May 2025

·

24 Reads

Despite growing interest in automated hate speech detection, most existing approaches overlook the linguistic diversity of online content. Multilingual instruction-tuned large language models such as LLaMA, Aya, Qwen, and BloomZ offer promising capabilities across languages, but their effectiveness in identifying hate speech through zero-shot and few-shot prompting remains underexplored. This work evaluates LLM prompting-based detection across eight non-English languages, utilizing several prompting techniques and comparing them to fine-tuned encoder models. We show that while zero-shot and few-shot prompting lag behind fine-tuned encoder models on most of the real-world evaluation sets, they achieve better generalization on functional tests for hate speech detection. Our study also reveals that prompt design plays a critical role, with each language often requiring customized prompting techniques to maximize performance.


Figure 1: Scatter plots of eight features extracted from a Chinese corpus during the data cleaning process, with data points color-coded according to their anomaly labels. The yellow points represent high-quality data, while the purple points indicate low-quality data.
Figure 3: Compare DCAD-2000 with other multilingual corpora.
DCAD-2000: A Multilingual Dataset across 2000+ Languages with Data Cleaning as Anomaly Detection

February 2025

·

10 Reads

The rapid development of multilingual large language models (LLMs) highlights the need for high-quality, diverse, and clean multilingual datasets. In this paper, we introduce DCAD-2000 (Data Cleaning as Anomaly Detection), a large-scale multilingual corpus built using newly extracted Common Crawl data and existing multilingual datasets. DCAD-2000 includes over 2,282 languages, 46.72TB of data, and 8.63 billion documents, spanning 155 high- and medium-resource languages and 159 writing scripts. To overcome the limitations of current data cleaning methods, which rely on manual heuristic thresholds, we propose reframing data cleaning as an anomaly detection task. This dynamic filtering approach significantly enhances data quality by identifying and removing noisy or anomalous content. We evaluate the quality of DCAD-2000 on the FineTask benchmark, demonstrating substantial improvements in multilingual dataset quality and task performance.


Beyond Literal Token Overlap: Token Alignability for Multilinguality

February 2025

·

8 Reads

Previous work has considered token overlap, or even similarity of token distributions, as predictors for multilinguality and cross-lingual knowledge transfer in language models. However, these very literal metrics assign large distances to language pairs with different scripts, which can nevertheless show good cross-linguality. This limits the explanatory strength of token overlap for knowledge transfer between language pairs that use distinct scripts or follow different orthographic conventions. In this paper, we propose subword token alignability as a new way to understand the impact and quality of multilingual tokenisation. In particular, this metric predicts multilinguality much better when scripts are disparate and the overlap of literal tokens is low. We analyse this metric in the context of both encoder and decoder models, look at data size as a potential distractor, and discuss how this insight may be applied to multilingual tokenisation in future work. We recommend our subword token alignability metric for identifying optimal language pairs for cross-lingual transfer, as well as to guide the construction of better multilingual tokenisers in the future. We publish our code and reproducibility details.


Joint Localization and Activation Editing for Low-Resource Fine-Tuning

February 2025

·

2 Reads

Parameter-efficient fine-tuning (PEFT) methods, such as LoRA, are commonly used to adapt LLMs. However, the effectiveness of standard PEFT methods is limited in low-resource scenarios with only a few hundred examples. Recent advances in interpretability research have inspired the emergence of activation editing techniques, which modify the activations of specific model components. These methods, due to their extremely small parameter counts, show promise for small datasets. However, their performance is highly dependent on identifying the correct modules to edit and often lacks stability across different datasets. In this paper, we propose Joint Localization and Activation Editing (JoLA), a method that jointly learns (1) which heads in the Transformer to edit (2) whether the intervention should be additive, multiplicative, or both and (3) the intervention parameters themselves - the vectors applied as additive offsets or multiplicative scalings to the head output. Through evaluations on three benchmarks spanning commonsense reasoning, natural language understanding, and natural language generation, we demonstrate that JoLA consistently outperforms existing methods.



Citations (53)


... Tonneau et al. (2024) disclose the intra-language geographical bias of English, Arabic, and Spanish hate speech datasets, as inferring the location of each tweet's author reveals that a handful of countries are disproportionately overrepresented in the datasets. Masud et al. (2024) examine LLMs' ability to represent diverse groups using persona-based attributes and geographical priming, finding that persona-based mimicry increases annotation variability, while geographical signals improve regional alignment, with implications for using LLMs as cost-effective proxies for underrepresented annotator demographics. ...

Reference:

Survey of Cultural Awareness in Language Models: Text and Beyond
Hate Personified: Investigating the role of LLMs in content moderation
  • Citing Conference Paper
  • January 2024

... LLMen FFN geruzetako neuronetan jasota dagoen jakintza espezifikoa ikerketa-gai aktiboa da (Dai et al., 2022;Lai et al., 2024;Leng eta Xiong, 2024), eta hizkuntza bakoitzerako espezifikoak diren neuronak ereduan nola banatuta dauden jakiteko metodo bat ere badago (Tang et al., 2024). Metodo horretan, Hizkuntza Aktibaziorako Probabilitate Entropia (Language Activation Probability Entropy, LAPE) deituriko metrika kalkulatzen da ereduko neurona bakoitzerako, eta LAPE txikia duten eta hizkuntza jakin batekin aktibatzeko probabilitate handia duten neuronak hizkuntza horretako espezifikotzat hartzen dira. ...

Style-Specific Neurons for Steering LLMs in Text Style Transfer
  • Citing Conference Paper
  • January 2024

... While large language models (LLMs) have grown remarkably in recent years, several studies report that they still struggle with fine-grained characterlevel manipulations, such as inserting, deleting, or extracting individual characters within tokens (Edman et al., 2024;Wang et al., 2024;Chai et al., 2024;Shin and Kaneko, 2024). Although most LLMs operate over subword tokens, true mastery of subtoken information is essential for a range of applications, such as morphological inflection (Marco and Fraser, 2024), letter counting , typoglycemia , and handling typos (Tsuji et al., 2025). To improve their reliability in such scenarios, we must understand how LLMs internally represent and process characters. ...

Subword Segmentation in LLMs: Looking at Inflection and Consistency
  • Citing Conference Paper
  • January 2024

... We use sequence packing for dataloading during training along with FSDP. We evaluate both the 7.5B BPE baseline and AU-Net 2 on the CUTE benchmark Edman et al. (2024), which tests a model's ability to manipulate both words and characters. As shown in Table 7, our byte-level model performs better on character-level tasks, while the BPE baseline takes the lead on word-level ones. ...

CUTE: Measuring LLMs’ Understanding of Their Tokens
  • Citing Conference Paper
  • January 2024

... Jalili Sabet et al. (2020) for crosslingual word alignment. Similarly to Hämmerl et al. (2024), we experiment with strategies that result in a range from weak to strong alignment, where strategies for weaker alignments typically allow a higher range of semantic similarity and multiple possible alignments, while strategies for stronger alignments are more restrictive towards a high semantic similarity and may only include one good alignment (Figure 1). ...

Understanding Cross-Lingual Alignment—A Survey

... Cross-domain transfer occurs when item metadata is reformulated into coherent sentences serving as inputs to pretrained LLMs [124,125]. The use of cross-lingual feedback and multilingual datasets like Amazon-M2 enhances multilingual recommendation [126,127]. The degree of difficulty posed by integrating information uniformly across languages is tackled by recent alignment approaches [128]. ...

LLMs Beyond English: Scaling the Multilingual Capability of LLMs with Cross-Lingual Feedback
  • Citing Conference Paper
  • January 2024

... Johnson and Khoshgoftaar (2019) provided a comprehensive survey of techniques for handling imbalanced data in machine learning, several of which we incorporated in our approach. More specific to NLP, Henning et al. (2023) explored techniques for addressing class imbalance in transformer-based text classification, demonstrating that appropriate loss functions and sampling strategies can significantly improve performance for minority classes. Our test-time augmentation approach (see Section 4.3) draws on work by Shanmugam et al. (2021), who demonstrated that augmented inference can improve classification performance, particularly in challenging examples. ...

A Survey of Methods for Addressing Class Imbalance in Deep-Learning Based Natural Language Processing
  • Citing Conference Paper
  • January 2023

... Weight-space fusion methods merge independently fine-tuned adapters through averaging, clustering, or simple arithmetic. AdapterSoup [19] trains a set of domain-specific adapters and selects which ones to average at test time for each novel domain. LoRA Soups [56] demonstrate that concatenating LoRAs using optimally learned weights outperforms traditional model and data-merging techniques. ...

AdapterSoup: Weight Averaging to Improve Generalization of Pretrained Language Models
  • Citing Conference Paper
  • January 2023

... A key challenge for multilingual models is task interference, where too many languages tend to degrade model performance (Zaremoodi, Buntine, and Haffari 2018;Wang, Lipton, and Tsvetkov 2020). As a result, research in this direction has been alleviating such interference by developing various parameter-separation schemes (Baziotis et al. 2022;Chronopoulou, Stojanovski, and Fraser 2023) and using gradient-based methods to update language-specific pa-rameters (Wang and Zhang 2022;He et al. 2023). In our work, we use a standard Transformer model following Johnson et al. (2017) and Liu et al. (2021). ...

Language-Family Adapters for Low-Resource Multilingual Neural Machine Translation
  • Citing Conference Paper
  • January 2023

... The review emphasizes practical constraints such as model complexity, linguistic variation, and lack of labeled data, positioning transfer learning as the most scalable solution for inclusive NLP development (Gunnam, 2022). Hangya et al. (2022) propose an unsupervised technique to enhance multilingual model performance for LRLs using word translation bootstrapping. Their method improves cross-lingual alignment without parallel corpora by leveraging monolingual embeddings and unsupervised word translation to better encode LRL vocabulary. ...

Improving Low-Resource Languages in Pre-Trained Multilingual Language Models
  • Citing Conference Paper
  • January 2022