Dominik Macko’s scientific contributions

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (11)


Figure 1: An overview of mdok approach for binary (top) and multiclass (bottom) detection of machine-generated text.
Comparison of Various Models and Variants on Validation Set of Subtask 1
Comparison of Various Models and Variants on Validation Set of Subtask 2
Official Leaderboard of Subtask 2 (for Teams Ourperforming the Baseline)
mdok of KInIT: Robustly Fine-tuned LLM for Binary and Multiclass AI-Generated Text Detection
  • Preprint
  • File available

June 2025

·

5 Reads

Dominik Macko

The large language models (LLMs) are able to generate high-quality texts in multiple languages. Such texts are often not recognizable by humans as generated, and therefore present a potential of LLMs for misuse (e.g., plagiarism, spams, disinformation spreading). An automated detection is able to assist humans to indicate the machine-generated texts; however, its robustness to out-of-distribution data is still challenging. This notebook describes our mdok approach in robust detection, based on fine-tuning smaller LLMs for text classification. It is applied to both subtasks of Voight-Kampff Generative AI Detection 2025, providing remarkable performance in binary detection as well as in multiclass (1st rank) classification of various cases of human-AI collaboration.

Download

Figure 1. Per-year Mean Score of the two fine-tuned detectors for MultiClaim texts. Both detectors independently show increasing Mean Scores in recent years.
Figure 2. Per-year proportion of the MultiClaim texts detected to be machine-generated. Proportion is increasing in recent years, with the highest increase in 2023 (after ChatGPT release).
Figure 4. Proportion of the texts detected to be machine-generated in different datasets.
Figure 5. The per-month proportion of the texts detected to be machine-generated in the USC-X dataset for the election year of 2024 (all languages on the left, English texts only on the right). Proportion is increasing towards the election date in November 2024.
Overview of the used datasets. The upper part of the table includes the datasets that we have
Beyond speculation: Measuring the growing presence of LLM-generated texts in multilingual disinformation

March 2025

·

27 Reads

Dominik Macko

·

Aashish Anantha Ramakrishnan

·

·

[...]

·

Dongwon Lee

Increased sophistication of large language models (LLMs) and the consequent quality of generated multilingual text raises concerns about potential disinformation misuse. While humans struggle to distinguish LLM-generated content from human-written texts, the scholarly debate about their impact remains divided. Some argue that heightened fears are overblown due to natural ecosystem limitations, while others contend that specific "longtail" contexts face overlooked risks. Our study bridges this debate by providing the first empirical evidence of LLM presence in the latest real-world disinformation datasets, documenting the increase of machine-generated content following ChatGPT's release, and revealing crucial patterns across languages, platforms, and time periods.


Figure 2: Receiver operating characteristic curves of the detectors using the MultiSocial dataset.
Figure 3: Receiver operating characteristic curves of the detectors using the MIX dataset.
Evaluation of detection performance on in- distribution data. The highest value in each column is boldfaced.
Per-language AUC ROC performance of the detectors on in-distribution data. The highest value in each dataset per each language is boldfaced.
Increasing the Robustness of the Fine-tuned Multilingual Machine-Generated Text Detectors

March 2025

·

4 Reads

Since the proliferation of LLMs, there have been concerns about their misuse for harmful content creation and spreading. Recent studies justify such fears, providing evidence of LLM vulnerabilities and high potential of their misuse. Humans are no longer able to distinguish between high-quality machine-generated and authentic human-written texts. Therefore, it is crucial to develop automated means to accurately detect machine-generated content. It would enable to identify such content in online information space, thus providing an additional information about its credibility. This work addresses the problem by proposing a robust fine-tuning process of LLMs for the detection task, making the detectors more robust against obfuscation and more generalizable to out-of-distribution data.


Figure 5: Distribution of meta-evaluation scores assigned by individual meta-evaluators. Text counts are for all generators combined.
Figure 13: Receiver operating characteristic curves of the selected detection methods.
Evaluation of LLM Vulnerabilities to Being Misused for Personalized Disinformation Generation

December 2024

·

46 Reads

·

1 Citation

The capabilities of recent large language models (LLMs) to generate high-quality content indistinguishable by humans from human-written texts rises many concerns regarding their misuse. Previous research has shown that LLMs can be effectively misused for generating disinformation news articles following predefined narratives. Their capabilities to generate personalized (in various aspects) content have also been evaluated and mostly found usable. However, a combination of personalization and disinformation abilities of LLMs has not been comprehensively studied yet. Such a dangerous combination should trigger integrated safety filters of the LLMs, if there are some. This study fills this gap by evaluation of vulnerabilities of recent open and closed LLMs, and their willingness to generate personalized disinformation news articles in English. We further explore whether the LLMs can reliably meta-evaluate the personalization quality and whether the personalization affects the generated-texts detectability. Our results demonstrate the need for stronger safety-filters and disclaimers, as those are not properly functioning in most of the evaluated LLMs. Additionally, our study revealed that the personalization actually reduces the safety-filter activations; thus effectively functioning as a jailbreak. Such behavior must be urgently addressed by LLM developers and service providers.


Figure 1: MultiSocial coverage of languages.
Per-platform AUC ROC performance of fine-tuned MGT detectors category. N/A refers to not enough samples per each class (at least 10) making AUC ROC value irrelevant. Platform [AUC ROC] Detector Generator Discord Gab Telegram Twitter WhatsApp all
MultiSocial: Multilingual Benchmark of Machine-Generated Text Detection of Social-Media Texts

June 2024

·

67 Reads

Recent LLMs are able to generate high-quality multilingual texts, indistinguishable for humans from authentic human-written ones. Research in machine-generated text detection is however mostly focused on the English language and longer texts, such as news articles, scientific papers or student essays. Social-media texts are usually much shorter and often feature informal language, grammatical errors, or distinct linguistic items (e.g., emoticons, hashtags). There is a gap in studying the ability of existing methods in detection of such texts, reflected also in the lack of existing multilingual benchmark datasets. To fill this gap we propose the first multilingual (22 languages) and multi-platform (5 social media platforms) dataset for benchmarking machine-generated text detection in the social-media domain, called MultiSocial. It contains 472,097 texts, of which about 58k are human-written and approximately the same amount is generated by each of 7 multilingual LLMs. We use this benchmark to compare existing detection methods in zero-shot as well as fine-tuned form. Our results indicate that the fine-tuned detectors have no problem to be trained on social-media texts and that the platform selection for training matters.







Citations (6)


... The robustness of machine-generated text detection methods against authorship obfuscation methods has been explored in [16]. Although it has been focused on multilingual settings and the results differ among languages, it has shown that fine-tuned detection methods are more robust against obfuscation than the statistical methods, while offering significantly higher detection performance. ...

Reference:

mdok of KInIT: Robustly Fine-tuned LLM for Binary and Multiclass AI-Generated Text Detection
Authorship Obfuscation in Multilingual Machine-Generated Text Detection
  • Citing Conference Paper
  • January 2024

... AI-generated articles, deepfakes, and synthetically created social media posts can now be used for disinformation campaigns [35]. The issue is amplified by the increasing amount of false information, with false * -Equal contribution claims spreading faster than truthful ones, particularly on social media platforms [40]. This spread is enhanced by recommendation algorithms that prioritize engagement over accuracy, undermining public trust in displayed content [35]. ...

Disinformation Capabilities of Large Language Models
  • Citing Conference Paper
  • January 2024

... Traditional approaches, reliant on human annotators, are resource-intensive and may struggle to scale as new forms of harmful language emerge (Vetagiri et al., 2024). Meanwhile, recent advances in Large Language Models (LLMs) have demonstrated impressive abilities in text generation, paraphrasing, and nuanced understanding of linguistic context (Kurt Pehlivanoglu et al., 2024;Wuraola et al., 2024;Tripto et al., 2024). ...

A Ship of Theseus: Curious Cases of Paraphrasing in LLM-Generated Texts
  • Citing Conference Paper
  • January 2024

... A binary machine-generated text detection is a well researched task, typically addressed by stylometric methods (e.g., a machine learning classifier trained on TF-IDF features), statistical methods (e.g., utilizing perplexity, entropy, or likelihood) [8,9], or fine-tuned language models for classification task (e.g., by supervised or contrastive learning) [10,11]. Most of the detection methods can be directly applied by existing frameworks, such as MGTBench [12] or IMGTB [13]. A multiclass machine-generated text classification is mostly researched in related authorship attribution task, identifying of the author (generator) of the text. ...

IMGTB: A Framework for Machine-Generated Text Detection Benchmarking

... A binary machine-generated text detection is a well researched task, typically addressed by stylometric methods (e.g., a machine learning classifier trained on TF-IDF features), statistical methods (e.g., utilizing perplexity, entropy, or likelihood) [8,9], or fine-tuned language models for classification task (e.g., by supervised or contrastive learning) [10,11]. Most of the detection methods can be directly applied by existing frameworks, such as MGTBench [12] or IMGTB [13]. ...

KInIT at SemEval-2024 Task 8: Fine-tuned LLMs for Multilingual Machine-Generated Text Detection

... Such approach has been proposed in [6] and shown to increase even the generalization to out-of-distribution data. The training data mixture consists of social-media texts from the MultiSocial [17] dataset, news articles from the MULTITuDE [18] dataset, and obfuscated texts from [16]. For validation (i.e., model-checkpoint selection), the approach uses a unique massively multi-generator (75 generators) multilingual (7 languages) data of MIX2k composition of 18 existing labeled datasets to represent out-of-distribution data. ...

MULTITuDE: Large-Scale Multilingual Machine-Generated Text Detection Benchmark