Jörg Tiedemann’s scientific contributions

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (6)


Overview of the different objectives considered in this study. Top two rows: two-stacks (encoder-decoder) models; bottom three rows: single-stack (encoder-only or decoder-only) models.
Two Stacks Are Better Than One: A Comparison of Language Modeling and Translation as Multilingual Pretraining Objectives
  • Preprint
  • File available

July 2024

·

10 Reads

Zihao Li

·

Shaoxiong Ji

·

·

[...]

·

Jörg Tiedemann

Pretrained language models (PLMs) display impressive performances and have captured the attention of the NLP community. Establishing the best practices in pretraining has therefore become a major point of focus for much of NLP research -- especially since the insights developed for monolingual English models need not carry to more complex multilingual. One significant caveat of the current state of the art is that different works are rarely comparable: they often discuss different parameter counts, training data, and evaluation methodology. This paper proposes a comparison of multilingual pretraining objectives in a controlled methodological environment. We ensure that training data and model architectures are comparable, and discuss the downstream performances across 6 languages that we observe in probing and fine-tuning scenarios. We make two key observations: (1) the architecture dictates which pretraining objective is optimal; (2) multilingual translation is a very effective pre-training objective under the right conditions. We make our code, data, and model weights available at \texttt{\url{https://github.com/Helsinki-NLP/lm-vs-mt}}.

Download



Figure 1: Median number of tokens in the datasets
A summary of datasets. Note we hold out a portion of the original training set as the validation set if the original dataset does not contain a validation set.
Domain-specific Continued Pretraining of Language Models for Capturing Long Context in Mental Health

April 2023

·

135 Reads

Pretrained language models have been used in various natural language processing applications. In the mental health domain, domain-specific language models are pretrained and released, which facilitates the early detection of mental health conditions. Social posts, e.g., on Reddit, are usually long documents. However, there are no domain-specific pretrained models for long-sequence modeling in the mental health domain. This paper conducts domain-specific continued pretraining to capture the long context for mental health. Specifically, we train and release MentalXLNet and MentalLongformer based on XLNet and Longformer. We evaluate the mental health classification performance and the long-range ability of these two domain-specific pretrained models. Our models are released in HuggingFace.


Figure 1: An example of a humorous segment that preceded laughter
Figure 2: Joey entering the room wearing many layers of clothes provoked audience laughter non-verbally
Figure 3: Joey inquiring whether Monica had cooked a person after tasting her food.
Figure 4: Phoebe having a grumpy facial expression while delivering a laughter provoking line.
Figure 5: Descriptions produced by the image captioning model
When to Laugh and How Hard? A Multimodal Approach to Detecting Humor and its Intensity

November 2022

·

263 Reads

·

4 Citations

Prerecorded laughter accompanying dialog in comedy TV shows encourages the audience to laugh by clearly marking humorous moments in the show. We present an approach for automatically detecting humor in the Friends TV show using multimodal data. Our model is capable of recognizing whether an utterance is humorous or not and assess the intensity of it. We use the prerecorded laughter in the show as annotation as it marks humor and the length of the audience's laughter tells us how funny a given joke is. We evaluate the model on episodes the model has not been exposed to during the training phase. Our results show that the model is capable of correctly detecting whether an utterance is humorous 78% of the time and how long the audience's laughter reaction should last with a mean absolute error of 600 milliseconds .


Figure 1: An example of a humorous segment that preceded laughter
Figure 2: Joey entering the room wearing many layers of clothes provoked audience laughter non-verbally
Figure 3: Joey inquiring whether Monica had cooked a person after tasting her food.
Figure 4: Phoebe having a grumpy facial expression while delivering a laughter provoking line.
Figure 5: Descriptions produced by the image captioning model
When to Laugh and How Hard? A Multimodal Approach to Detecting Humor and its Intensity

November 2022

·

136 Reads

Prerecorded laughter accompanying dialog in comedy TV shows encourages the audience to laugh by clearly marking humorous moments in the show. We present an approach for automatically detecting humor in the Friends TV show using multimodal data. Our model is capable of recognizing whether an utterance is humorous or not and assess the intensity of it. We use the prerecorded laughter in the show as annotation as it marks humor and the length of the audience's laughter tells us how funny a given joke is. We evaluate the model on episodes the model has not been exposed to during the training phase. Our results show that the model is capable of correctly detecting whether an utterance is humorous 78% of the time and how long the audience's laughter reaction should last with a mean absolute error of 600 milliseconds.

Citations (2)


... Helsinki-NLP. This team submitted models (de Gibert et al., 2024) exclusively to the unconstrained open track. Alongside the data provided in the task, such as PILAR, they utilized additional monolingual resources like Wikipedia dumps and dictionary definitions. ...

Reference:

Findings of the WMT 2024 Shared Task Translation into Low-Resource Languages of Spain: Blending Rule-Based and Neural Systems
Hybrid Distillation from RBMT and NMT: Helsinki-NLP’s Submission to the Shared Task on Translation into Low-Resource Languages of Spain
  • Citing Conference Paper
  • January 2024

... Regarding facial expression, Duchenne smiles, which are generally considered as the smile of genuine enjoyment, are usually accompanied by both the raising of cheeks and crow's feet around the eyes, while non-Duchenne smiles do not include the activation of muscles associated with "happy eyes" [10,11]. In the acted Duchenne smiles, other facial actions, for example, eyebrow rising, nose wrinkling and lip stretcher occur occasionally [12]. Moreover, some works investigated the facial expressions correlated with the perceived degree of intensity [13,14]. ...

When to Laugh and How Hard? A Multimodal Approach to Detecting Humor and its Intensity