Abhik Bhattacharjee’s scientific contributions

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (17)


Performance of Various Tasks and Abilities across Languages and Cultural Contexts (D: Direct/, C: Cultural) for GPT-3.5-Turbo Lang UOT SIT PST FBT AST HTT SST FRT EMO DES INT KNO BEL NLC
Performance of Various Tasks and Abilities across Languages and Cultural Contexts (D: Direct/, C: Cultural) for Llama-3.1-8b-Instruct Lang UOT SIT PST FBT AST HTT SST FRT EMO DES INT KNO BEL NLC
Multi-ToM: Evaluating Multilingual Theory of Mind Capabilities in Large Language Models
  • Preprint
  • File available

November 2024

·

19 Reads

Jayanta Sadhu

·

Ayan Antik Khan

·

Noshin Nawal

·

[...]

·

Rifat Shahriyar

Theory of Mind (ToM) refers to the cognitive ability to infer and attribute mental states to oneself and others. As large language models (LLMs) are increasingly evaluated for social and cognitive capabilities, it remains unclear to what extent these models demonstrate ToM across diverse languages and cultural contexts. In this paper, we introduce a comprehensive study of multilingual ToM capabilities aimed at addressing this gap. Our approach includes two key components: (1) We translate existing ToM datasets into multiple languages, effectively creating a multilingual ToM dataset and (2) We enrich these translations with culturally specific elements to reflect the social and cognitive scenarios relevant to diverse populations. We conduct extensive evaluations of six state-of-the-art LLMs to measure their ToM performance across both the translated and culturally adapted datasets. The results highlight the influence of linguistic and cultural diversity on the models' ability to exhibit ToM, and questions their social reasoning capabilities. This work lays the groundwork for future research into enhancing LLMs' cross-cultural social cognition and contributes to the development of more culturally aware and socially intelligent AI systems. All our data and code are publicly available.

Download

Figure 3: Male vs Female terms used for aggregation
Sentence structures for contextual bias
An Empirical Study on the Characteristics of Bias upon Context Length Variation for Bangla

June 2024

·

62 Reads

Pretrained language models inherently exhibit various social biases, prompting a crucial examination of their social impact across various linguistic contexts due to their widespread usage. Previous studies have provided numerous methods for intrinsic bias measurements, predominantly focused on high-resource languages. In this work, we aim to extend these investigations to Bangla, a low-resource language. Specifically, in this study, we (1) create a dataset for intrinsic gender bias measurement in Bangla, (2) discuss necessary adaptations to apply existing bias measurement methods for Bangla, and (3) examine the impact of context length variation on bias measurement, a factor that has been overlooked in previous studies. Through our experiments, we demonstrate a clear dependency of bias metrics on context length, highlighting the need for nuanced considerations in Bangla bias analysis. We consider our work as a stepping stone for bias measurement in the Bangla Language and make all of our resources publicly available to support future research.





BanglaParaphrase: A High-Quality Bangla Paraphrase Dataset

October 2022

·

56 Reads

In this work, we present BanglaParaphrase, a high-quality synthetic Bangla Paraphrase dataset curated by a novel filtering pipeline. We aim to take a step towards alleviating the low resource status of the Bangla language in the NLP domain through the introduction of BanglaParaphrase, which ensures quality by preserving both semantics and diversity, making it particularly useful to enhance other Bangla datasets. We show a detailed comparative analysis between our dataset and models trained on it with other existing works to establish the viability of our synthetic paraphrase data generation pipeline. We are making the dataset and models publicly available at https://github.com/csebuetnlp/banglaparaphrase to further the state of Bangla NLP.


GEMv2: Multilingual NLG Benchmarking in a Single Line of Code

June 2022

·

143 Reads

Evaluation in machine learning is usually informed by past choices, for example which datasets or metrics to use. This standardization enables the comparison on equal footing using leaderboards, but the evaluation choices become sub-optimal as better alternatives arise. This problem is especially pertinent in natural language generation which requires ever-improving suites of datasets, metrics, and human evaluation to make definitive claims. To make following best model evaluation practices easier, we introduce GEMv2. The new version of the Generation, Evaluation, and Metrics Benchmark introduces a modular infrastructure for dataset, model, and metric developers to benefit from each others work. GEMv2 supports 40 documented datasets in 51 languages. Models for all datasets can be evaluated online and our interactive data card creation and rendering tools make it easier to add new datasets to the living benchmark.


BanglaNLG: Benchmarks and Resources for Evaluating Low-Resource Natural Language Generation in Bangla

May 2022

·

117 Reads

This work presents BanglaNLG, a comprehensive benchmark for evaluating natural language generation (NLG) models in Bangla, a widely spoken yet low-resource language in the web domain. We aggregate three challenging conditional text generation tasks under the BanglaNLG benchmark. Then, using a clean corpus of 27.5 GB of Bangla data, we pretrain BanglaT5, a sequence-to-sequence Transformer model for Bangla. BanglaT5 achieves state-of-the-art performance in all of these tasks, outperforming mT5 (base) by up to 5.4%. We are making the BanglaT5 language model and a leaderboard publicly available in the hope of advancing future research and evaluation on Bangla NLG. The resources can be found at https://github.com/csebuetnlp/BanglaNLG.


(We mention the upper and lower values of the different tasks for each model)
BanglaBERT: Language Model Pretraining and Benchmarks for Low-Resource Language Understanding Evaluation in Bangla

May 2022

·

263 Reads

·

124 Citations

In this work, we introduce BanglaBERT, a BERT-based Natural Language Understanding (NLU) model pretrained in Bangla, a widely spoken yet low-resource language in the NLP literature. To pretrain BanglaBERT, we collect 27.5 GB of Bangla pretraining data (dubbed 'Bangla2B+') by crawling 110 popular Bangla sites. We introduce two downstream task datasets on natural language inference and question answering and benchmark on four diverse NLU tasks covering text classification, sequence labeling, and span prediction. In the process, we bring them under the first-ever Bangla Language Understanding Benchmark (BLUB). BanglaBERT achieves state-of-the-art results outperforming multilingual and monolingual models. We are making the models, datasets, and a leader-board publicly available at https://github. com/csebuetnlp/banglabert to advance Bangla NLP.



Citations (9)


... They also discussed the rational of including emoji's and stopwords in Bangla since they may alter the meaning of the text due to the complex nature of the Bangla Language. Similar studies in Bangla NLP includes BanglaParaphrase by Akil et al. (2022) where they synthetically generated a high-quality corpus focused on lexical diversity and semantic similarity. A pre-trained BanglaT5 model, fine-tuned on this dataset, is also available, making it a valuable tool for paraphrasingbased augmentation. ...

Reference:

BDA: Bangla Text Data Augmentation Framework
BanglaParaphrase: A High-Quality Bangla Paraphrase Dataset
  • Citing Conference Paper
  • January 2022

... Works of Bhattacharjee et al. (2022);Hasan et al. (2021Hasan et al. ( , 2020Akil et al. (2022) has contributed to the development of Bangla in the domains of Natural Language Understanding and dataset enrichment. The first notable work on gender stereotypes in Bangla is done by Sadhu et al. (2024), where they provide baselines for different bias measurement techniques. Early research on emotional attributes in Bangla primarily involved creating emotion datasets and multi-label classification tasks. ...

An Empirical Study on the Characteristics of Bias upon Context Length Variation for Bangla
  • Citing Conference Paper
  • January 2024

... Hugging Face link Parameters Pretrained on Bangala-T5 (Bhattacharjee et al., 2023) https://huggingface.co/csebuetnlp/BanglaT5 247M Bengali2B+ mT0-base (Muennighoff et al., 2023) https://huggingface.co/bigscience/mt0-base 582M mC4 mT5-Base (Xue et al., 2021) https://huggingface.co/google/mt5-base 582M mC4 mBART-50 (Tang et al., 2021) https://huggingface.co/facebook/mbart-large-50 610M CC25 Table 9 The hyperparameter search space used in tuning and the selected optimal hyperparameters for each model. ...

BanglaNLG and BanglaT5: Benchmarks and Resources for Evaluating Low-Resource Natural Language Generation in Bangla
  • Citing Conference Paper
  • January 2023

... Larger models such as GPT-3 excel at surface-level generation but remain limited in causal reasoning, indicating architectural constraints in effectively modelling semantics and world knowledge (Holmes and Tuomi 2022). Reviewing thirdparty experiments that stress test models against adversarial examples and common fallacies reveals robustness issues stemming from architectural design tradeoffs (Gehrmann et al. 2022). Being attuned to cutting-edge research around architectural improvements also surfaces current limitations in state-of-the-art generative models (Fuhr and Sumpter 2022). ...

GEMv2: Multilingual NLG Benchmarking in a Single Line of Code
  • Citing Conference Paper
  • January 2022

... • CrossSum [51] is a model fine-tuned based on mT5 using a multi-stage sampling strategy. The fine-tuning dataset used is CrossSum, a cross-lingual summary dataset that does not rely on English as the pivot language. ...

CrossSum: Beyond English-Centric Cross-Lingual Summarization for 1,500+ Language Pairs
  • Citing Conference Paper
  • January 2023

... Recent studies, including Hori et al. (2017), Yang et al. (2019), and Alam et al. (2021), indicate that combining visual and textual data improves performance in complex NLP tasks. For multimodal feature representation, we applied a feature fusion approach Nojavanasghari et al. (2016), integrating both visual and textual models such as Bangla-BERT (Sarker, 2020;Bhattacharjee et al., 2022). We added a dense layer with 100 neurons to each modality, then concatenated their outputs to cre-ate a unified feature representation, followed by a dense layer with 32 neurons and a sigmoid layer for classification. ...

BanglaBERT: Language Model Pretraining and Benchmarks for Low-Resource Language Understanding Evaluation in Bangla

... 4. Generative tasks: We evaluate model performance in machine translation and summarization on FLORES-200 [NLLB- Team et al., 2022] and XLSum [Hasan et al., 2021] respectively. For FLORES, we use all 21 languages (X ↔ English) and for XLSum, we use 15 languages based on language coverage of Aya 23 models. ...

XL-Sum: Large-Scale Multilingual Abstractive Summarization for 44 Languages

... Preprocessing is a critical phase that prepares raw text data for input into generative models, ensuring the data is clean, consistent, and ready for effective processing. It involves two main tasks: • Text Normalization: This step uses the BUET normalizer (Hasan et al., 2020) to standardize characters with Unicode NFKC. Nontextual elements like URLs and emojis are removed, excessive whitespace is managed, and redundant punctuation characters are reduced. ...

Not Low-Resource Anymore: Aligner Ensembling, Batch Filtering, and New Datasets for Bengali-English Machine Translation