Kosuke Yamada’s research while affiliated with Nagoya University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (18)


Transformer-based Live Update Generation for Soccer Matches from Microblog PostsTransformer モデルを利用したマイクロブログからのサッカー速報生成
  • Article
  • Full-text available

December 2024

·

6 Reads

Journal of Natural Language Processing

Masashi Oshika

·

Kosuke Yamada

·

Ryohei Sasano

·

When a sports match is broadcast, X users often enjoy sharing the comment and it is possible to roughly understand a match’s progress by reading these posts. However, because of the diverse nature of posts, it can be challenging to quickly grasp a match’s progress. In this study, we focus on soccer matches and work on building a system to generate live updates from posts so that users can instantly grasp a match’s progress. Our system is based on a large language model T5, and outputs updates at certain times by inputting posts related to a specific match. However simply applying the model to this task caused two problems of the number of generated updates and redundant updates. Therefore, we propose mechanisms that incorporate a classifier to control the number of generated updates and a mechanism that takes into account the previous updates to mitigate redundancy.

Download

Semantic Frame Induction with Deep Metric Learning深層距離学習を用いた動詞の意味フレーム推定

December 2023

·

23 Reads

Journal of Natural Language Processing

Recent studies have demonstrated the usefulness of contextualized word embeddings in unsupervised semantic frame induction. However, they have also revealed that generic contextualized embeddings are not always consistent with human intuitions about semantic frames, which causes unsatisfactory performance for frame induction based on contextualized embeddings. In this paper, we tackle supervised semantic frame induction, which assumes the existence of frame-annotated data for a subset of verbs in a corpus and propose to fine-tune contextualized word embedding models using deep metric learning for high-performance semantic frame induction methods. Our experiments on FrameNet show that fine-tuning with deep metric learning considerably improves the clustering evaluation scores by about 8 points or more. We also demonstrate that our approach is effective even when the number of training instances is small.


Acquiring Frame Element Knowledge with Deep Metric Learning for Semantic Frame Induction

May 2023

·

4 Reads

The semantic frame induction tasks are defined as a clustering of words into the frames that they evoke, and a clustering of their arguments according to the frame element roles that they should fill. In this paper, we address the latter task of argument clustering, which aims to acquire frame element knowledge, and propose a method that applies deep metric learning. In this method, a pre-trained language model is fine-tuned to be suitable for distinguishing frame element roles through the use of frame-annotated data, and argument clustering is performed with embeddings obtained from the fine-tuned model. Experimental results on FrameNet demonstrate that our method achieves substantially better performance than existing methods.


Semantic Frame Induction with Deep Metric Learning

April 2023

·

8 Reads

Recent studies have demonstrated the usefulness of contextualized word embeddings in unsupervised semantic frame induction. However, they have also revealed that generic contextualized embeddings are not always consistent with human intuitions about semantic frames, which causes unsatisfactory performance for frame induction based on contextualized embeddings. In this paper, we address supervised semantic frame induction, which assumes the existence of frame-annotated data for a subset of predicates in a corpus and aims to build a frame induction model that leverages the annotated data. We propose a model that uses deep metric learning to fine-tune a contextualized embedding model, and we apply the fine-tuned contextualized embeddings to perform semantic frame induction. Our experiments on FrameNet show that fine-tuning with deep metric learning considerably improves the clustering evaluation scores, namely, the B-cubed F-score and Purity F-score, by about 8 points or more. We also demonstrate that our approach is effective even when the number of training instances is small.






Semantic Frame Induction using Masked Word Embeddings and Two-Step Clusteringマスクされた単語埋め込みと 2 段階クラスタリングを用いた動詞の意味フレーム推定

June 2022

·

9 Reads

·

1 Citation

Journal of Natural Language Processing

Recent studies on semantic frame induction show that relatively high performance has been achieved by using clustering-based methods with contextualized word embeddings. However, there are two potential drawbacks to these methods: one is that they focus too much on the superficial information of the frame-evoking verb and the other is that they tend to divide the instances of the same verb into too many different frame clusters. To overcome these drawbacks, we propose a semantic frame induction method using masked word embeddings and two-step clustering. Through experiments on the dataset from the English FrameNet, we demonstrate that using the masked word embeddings is effective for avoiding too much reliance on the surface information of frame-evoking verbs and that two-step clustering can improve the number of resulting frame clusters for the instances of the same verb.


Transformer-based Lexically Constrained Headline Generation

September 2021

·

14 Reads

This paper explores a variant of automatic headline generation methods, where a generated headline is required to include a given phrase such as a company or a product name. Previous methods using Transformer-based models generate a headline including a given phrase by providing the encoder with additional information corresponding to the given phrase. However, these methods cannot always include the phrase in the generated headline. Inspired by previous RNN-based methods generating token sequences in backward and forward directions from the given phrase, we propose a simple Transformer-based method that guarantees to include the given phrase in the high-quality generated headline. We also consider a new headline generation strategy that takes advantage of the controllable generation order of Transformer. Our experiments with the Japanese News Corpus demonstrate that our methods, which are guaranteed to include the phrase in the generated headline, achieve ROUGE scores comparable to previous Transformer-based methods. We also show that our generation strategy performs better than previous strategies.


Citations (6)


... The mainstream methods are mainly based on contextualized word embeddings (Qasem-iZadeh et al., 2019;Yamada et al., 2021b) such as BERT (Devlin et al., 2019). These methods leverage the observation that words evoking the same semantic frame tend to appear in similar contexts, resulting in their embeddings being grouped into the same cluster (Yamada et al., 2021a(Yamada et al., , 2023. However, while the frame induction task provides clusters of frames, it lacks interpretability because definitions of these clusters are not provided. ...

Reference:

Definition Generation for Automatically Induced Semantic Frame
Semantic Frame Induction with Deep Metric Learning
  • Citing Conference Paper
  • January 2023

... In Li et al. [26], the authors propose to introduce a multi-head self-attention mechanism into news HG based on the Transformer decoder, and designed a decoding selection strategy integrating top-k, top-p and penalty mechanisms to select important semantic information and then generate news headlines. In Yamada et al. [27], the authors propose a Transformer-based Seq2BF model that alternates forward and backward decoding to generate headlines with a given phrase. In Bukhtiyarov et al. [28], the authors fine-tuned two Transformer-based pre-trained models, mBART and BertSumAbs, and achieved good results. ...

Transformer-based Lexically Constrained Headline Generation
  • Citing Conference Paper
  • January 2021

... The mainstream methods are mainly based on contextualized word embeddings (Qasem-iZadeh et al., 2019;Yamada et al., 2021b) such as BERT (Devlin et al., 2019). These methods leverage the observation that words evoking the same semantic frame tend to appear in similar contexts, resulting in their embeddings being grouped into the same cluster (Yamada et al., 2021a(Yamada et al., , 2023. However, while the frame induction task provides clusters of frames, it lacks interpretability because definitions of these clusters are not provided. ...

Semantic Frame Induction using Masked Word Embeddings and Two-Step Clustering
  • Citing Conference Paper
  • January 2021

... where N c is the number of clusters, and D the data dimensionality. The term (D + 2) accounts for the mean and covariance parameters, and subtracting 1 corrects for the constraints, ensuring the sum of parameters equals 1 for mixture weights (Murphy, 2012;Yamada et al., 2021). Table 5 presents the performance of various models on estimating the number of emotions and clusters in both datasets, using accuracy, Spearman's rank correlation coefficient (ρ), and root mean square error (RMSE) metrics. ...

Verb Sense Clustering using Contextualized Word Representations for Semantic Frame Induction
  • Citing Conference Paper
  • January 2021

... ↑ indicates overall improvement compared to the previous step, ↓ an overall decrease. The SOTA row lists the best performing approaches to date: 1) SciBERT + MLP by Cohan et al. (2019) 2) Neural Semi-Markov CRFs by Yamada et al. (2020) 3) Contrastive Training of BigBird by Laboulaye (2021) 4) (Hochreiter and Schmidhuber, 1997) to capture dependencies between sentences, (3) a linear layer to project from the BiLSTM's hidden dimension to the number of tags, and (4) a CRF (Lafferty et al., 2001) layer to model dependencies between tags. We compare two embedding models: The default all-mpnet-base-v2 (Song et al., 2020) with NV-Embed (7B parameters) (Lee et al., 2024), the best-performing model on the English MTEB (as of December 2024). ...

Sequential Span Classification with Neural Semi-Markov CRFs for Biomedical Abstracts
  • Citing Conference Paper
  • January 2020

... Personality manifests in the individual characteristics of behavior, cognition, and emotional patterns and has been extensively predicted from textual information 21,22 . In early studies, Pennebaker et al. 23 built the Essays dataset containing 2468 anonymous essays tagged with the authors' Big Five personality traits and analyzed the correlation between the linguistic styles of the authors and their personality traits. ...

Incorporating Textual Information on User Behavior for Personality Prediction
  • Citing Conference Paper
  • January 2019