Alessandro Bria’s research while affiliated with University of Cassino and Southern Lazio and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (61)


Proposed approach
Proposed NER approach. Panel 1) shows the corpus generation, including annotation and the pre-processing of the raw text (sentence detection & tokenization). Panel 2) shows the fine-tuning phase, whereas panel 3) the validation phase. Both 2) and 3) are carried out considering a 10 fold cross-validation experimental setup (10 fold CV black dotted box)
Histogram of entity types. On the y-axis we show the count (on the left) and the a-priori class probability (on the right) of each entity type. On the x-axis we show the various entity types. In addition to the histogram, we also display the Lorenz curve (in orange), which illustrates the distribution of entities in terms of their occurrences
Proposed architecture: The architecture utilizes token embeddings generated by the NER system before the classification layer. Each token embedding classified by the NER system as an entity (Entity Embedding) undergoes a weighting process through a token attentional layer. This produces a weighted average of the same embedding size, named as Sentence Embedding. The sentence embeddings derived from all sentences in a patient’s clinical reports, are then fed into a sentence attentional layer, which shares weights with the token attentional layer. The outcome is a weighted average vector, maintaining the original embedding size \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${d_E}$$\end{document}, referred to as the patient embedding \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${{\mathbf{x}}^{(i)}}$$\end{document}. The patient embedding is the input of the risk assessment network
Example of Attentional Map: within each sentence in patient clinical reports, only the words identified as entities by the NER system, highlighted in yellow, are aggregated into the sentence embedding. These sentences receive a score assigned by the second attention layer (sentence level attention), with higher scores depicted in shades of red and lower scores tending towards blue. The text was translated from Italian to English for presentation purposes

+5

Hierarchical embedding attention for overall survival prediction in lung cancer from unstructured EHRs
  • Article
  • Full-text available

April 2025

·

24 Reads

BMC Medical Informatics and Decision Making

Domenico Paolo

·

·

·

[...]

·

The automated processing of Electronic Health Records (EHRs) poses a significant challenge due to their unstructured nature, rich in valuable, yet disorganized information. Natural Language Processing (NLP), particularly Named Entity Recognition (NER), has been instrumental in extracting structured information from EHR data. However, existing literature primarly focuses on extracting handcrafted clinical features through NLP and NER methods without delving into their learned representations. In this work, we explore the untapped potential of these representations by considering their contextual richness and entity-specific information. Our proposed methodology extracts representations generated by a transformer-based NER model on EHRs data, combines them using a hierarchical attention mechanism, and employs the obtained enriched representation as input for a clinical prediction model. Specifically, this study addresses Overall Survival (OS) in Non-Small Cell Lung Cancer (NSCLC) using unstructured EHRs data collected from an Italian clinical centre encompassing 838 records from 231 lung cancer patients. Whilst our study is applied on EHRs written in Italian, it serves as use case to prove the effectiveness of extracting and employing high level textual representations that capture relevant information as named entities. Our methodology is interpretable because the hierarchical attention mechanism highlights the information in EHRs that the model considers the most crucial during the decision-making process. We validated this interpretability by measuring the agreement of domain experts on the importance assigned by the hierarchical attention mechanism to EHRs information through a questionnaire. Results demonstrate the effectiveness of our method, showcasing statistically significant improvements over traditional manually extracted clinical features.

Download

Latent Diffusion Autoencoders: Toward Efficient and Meaningful Unsupervised Representation Learning in Medical Imaging

April 2025

·

12 Reads

This study presents Latent Diffusion Autoencoder (LDAE), a novel encoder-decoder diffusion-based framework for efficient and meaningful unsupervised learning in medical imaging, focusing on Alzheimer disease (AD) using brain MR from the ADNI database as a case study. Unlike conventional diffusion autoencoders operating in image space, LDAE applies the diffusion process in a compressed latent representation, improving computational efficiency and making 3D medical imaging representation learning tractable. To validate the proposed approach, we explore two key hypotheses: (i) LDAE effectively captures meaningful semantic representations on 3D brain MR associated with AD and ageing, and (ii) LDAE achieves high-quality image generation and reconstruction while being computationally efficient. Experimental results support both hypotheses: (i) linear-probe evaluations demonstrate promising diagnostic performance for AD (ROC-AUC: 90%, ACC: 84%) and age prediction (MAE: 4.1 years, RMSE: 5.2 years); (ii) the learned semantic representations enable attribute manipulation, yielding anatomically plausible modifications; (iii) semantic interpolation experiments show strong reconstruction of missing scans, with SSIM of 0.969 (MSE: 0.0019) for a 6-month gap. Even for longer gaps (24 months), the model maintains robust performance (SSIM > 0.93, MSE < 0.004), indicating an ability to capture temporal progression trends; (iv) compared to conventional diffusion autoencoders, LDAE significantly increases inference throughput (20x faster) while also enhancing reconstruction quality. These findings position LDAE as a promising framework for scalable medical imaging applications, with the potential to serve as a foundation model for medical image analysis. Code available at https://github.com/GabrieleLozupone/LDAE



Fig. 1 Preprocessing Pipeline.
-Year OS distributions for LUNG1 and CLARO datasets.
Hyperparameters of the risk-assessment network.
Results of C td -index on CLARO. All backbone layers are frozen, except for the last one (feature extractor).
Computational complexity analysis of the experimented models.
Predicting Lung Cancer Survival with Attention-based CT Slices Combination

November 2024

·

24 Reads

Accurate prognosis of Non-Small Cell Lung Cancer (NSCLC) is crucial for enhancing patient care and treatment outcomes. Despite the advancements in deep learning, the task of overall survival prediction in NSCLC has not fully leveraged these techniques, yet. This study introduces a novel methodology for predicting 2-year overall survival (OS) in NSCLC patients using CT scans. Our approach integrates CT scan representations produced by EfficientNetB0 with a soft attention mechanism to identify the most relevant slices for survival risk prediction, which are then analyzed by a risk-assessment network. To validate our method and ensure reproducibility, we employed the public LUNG1 dataset and a smaller private dataset. Our approach was compared to a benchmark 3D network and two variants of our methodology: on the LUNG1 it outperformed the competitors achieving a mean Ctd-index of 0.584 over 10-fold cross-validation. On the LUNG1 we also demonstrated the adaptability of our method with 4 other 2D backbones replacing the EfficientNetB0, confirming that our mechanism of combining 2D slice representations to construct a 3D volume representation is more effective for OS prediction compared to a traditional 3D approach. Finally, we used transfer learning on the private dataset, showing that it can significantly enhance performance in limited data scenarios, increasing the Ctd-index by 0.076 compared to model without transfer learning.


Example of a sample from OMI-DB dataset: a original image; b cropped image; c mask of the original image; d four patches of 500 × 500
RetinaNet architecture with ResNet50, ResNet101, and ResNet152 backbones and ImageNet, COCO weights, and from scratch initializations
FROC curves of RetinaNet model with three types of initialization in the first approach: a ResNet50 with ImageNet weights, b ResNet50 with COCO weights, c ResNet50 from scratch, d ResNet101 with ImageNet weights, e ResNet101 with COCO weights, f ResNet101 from scratch, g ResNet152 with ImageNet weights; in the second approach: h ResNet50 with ImageNet weights, i ResNet50 with COCO weights, j ResNet50 from scratch, k ResNet101 with ImageNet weights, l ResNet101 with COCO weights, m ResNet101 from scratch, n ResNet152 with ImageNet weights, o FROC curves of models in the two approaches with the highest TPR parameter
ROC curve of models in the classification step with the AUC: a ResNet50+ImageNet, b ResNet50+COCO, c ResNet152+ImageNet, d ResNet152+ImageNet
Transfer learning in breast mass detection and classification

August 2024

·

72 Reads

·

1 Citation

Journal of Ambient Intelligence and Humanized Computing

Covid-19 infection influenced the screening test rate of breast cancer worldwide due to the quarantine measures, routine procedures reduction, and delay of early diagnosis, causing high mortality risk and severity of the disease. X-ray mammography is the gold standard for diagnosing early signs of breast cancer, and Artificial Intelligence enables the detection of suspicious lesions and classifying them in terms of malignancy. This paper aimed to investigate mass detection and classification in a large-scale OPTIMAM dataset with 6000 cases and extracted 3524 images with masses in the mammograms of the Hologic manufacturer. The methodology of the detection step is to train the RetinaNet architecture of ResNet50, ResNet101, and ResNet152 backbones with three types of initializations by ImageNet and COCO weights and from scratch. The dataset was pre-processed to generate two types of input with entire mammograms and patches, which are stated as the first and the second approaches. The results show that in the first approach, RetinaNet of ResNet50 backbone with ImageNet and COCO weights and ResNet152 with the same weights performed 0.91 True Positive Rate at 0.78 False Positive Per Image, respectively. In contrast, in the second approach, ResNet152 with ImageNet weights reached 0.88 TPR at 0.78 FPPI. In the classification step, the Transfer Learning approach was applied with fine-tuning by adding L2-regularization and class weights to balance class distribution in the datasets.




AXIAL: Attention-based eXplainability for Interpretable Alzheimer's Localized Diagnosis using 2D CNNs on 3D MRI brain scans

July 2024

·

100 Reads

·

1 Citation

This study presents an innovative method for Alzheimer's disease diagnosis using 3D MRI designed to enhance the explainability of model decisions. Our approach adopts a soft attention mechanism, enabling 2D CNNs to extract volumetric representations. At the same time, the importance of each slice in decision-making is learned, allowing the generation of a voxel-level attention map to produces an explainable MRI. To test our method and ensure the reproducibility of our results, we chose a standardized collection of MRI data from the Alzheimer's Disease Neuroimaging Initiative (ADNI). On this dataset, our method significantly outperforms state-of-the-art methods in (i) distinguishing AD from cognitive normal (CN) with an accuracy of 0.856 and Matthew's correlation coefficient (MCC) of 0.712, representing improvements of 2.4\% and 5.3\% respectively over the second-best, and (ii) in the prognostic task of discerning stable from progressive mild cognitive impairment (MCI) with an accuracy of 0.725 and MCC of 0.443, showing improvements of 10.2\% and 20.5\% respectively over the second-best. We achieved this prognostic result by adopting a double transfer learning strategy, which enhanced sensitivity to morphological changes and facilitated early-stage AD detection. With voxel-level precision, our method identified which specific areas are being paid attention to, identifying these predominant brain regions: the \emph{hippocampus}, the \emph{amygdala}, the \emph{parahippocampal}, and the \emph{inferior lateral ventricles}. All these areas are clinically associated with AD development. Furthermore, our approach consistently found the same AD-related areas across different cross-validation folds, proving its robustness and precision in highlighting areas that align closely with known pathological markers of the disease.


Figure 3: Overall framework of the proposed method working with triplet networks.
Figure 4: The MVSLAM pipeline integrates depth estimation, pose estimation, and 3D reconstruction modules to generate a continuously updated 3D map of the surgical environment from monocular endoscopic video frames.
Towards AI-driven Next Generation Personalized Healthcare and Well-being

In the last few years Artificial Intelligence (AI) is emerging as a game changer in many areas of society and, in particular, its integration in medicine heralds a transformative approach towards personalized healthcare and well-being, promising significant improvements in diagnostic precision, therapeutic outcomes, and patient care. Our research explores the cutting-edge realms of multimodal AI, resilient AI, and healthcare robotics, aiming to harness the synergy of diverse data modalities and advanced computational models to redefine healthcare paradigms. This multidisciplinary effort seeks to bridge technology and clinical practice, advancing AI-driven next generation personalized healthcare and well-being.


Exploring Negated Entites for Named Entity Recognition in Italian Lung Cancer Clinical Reports

May 2024

·

29 Reads

This paper explores the potential of leveraging electronic health records (EHRs) for personalized health research through the application of artificial intelligence (AI) techniques, specifically Named Entity Recognition (NER). By extracting crucial patient information from clinical texts, including diagnoses, medications, symptoms, and lab tests, AI facilitates the rapid identification of relevant data, paving the way for future care paradigms. The study focuses on Non-small cell lung cancer (NSCLC) in Italian clinical notes, introducing a novel set of 29 clinical entities that include both presence or absence (negation) of relevant information associated with NSCLC. Using a state-of-the-art model pretrained on Italian biomedical texts, we achieve promising results (average F1-score of 80.8%), demonstrating the feasibility of employing AI for extracting biomedical information in the Italian language.


Citations (37)


... This challenge has urged researchers to explore innovative methods, leading to the integration of machine learning (ML) mechanisms [3,4]. By incorporating ML, IDS have become more adaptive and capable of identifying both known and unknown forms of attacks [5]. However, their reliance on learned patterns make them vulnerable to adversarial attacks, where inputs are slightly manipulated in a way to trick the IDS model into misclassify malicious activities as benign [6]. ...

Reference:

Toward Realistic Adversarial Attacks in IDS: A Novel Feasibility Metric for Transferability
Machine Learning in Network Intrusion Detection: A Cross-Dataset Generalization Study

IEEE Access

... The work presented by [13] proposes a two-stage methodology for detecting and classifying breast masses in the OPTIMAM (OMI-DB) dataset. The author used RetinaNet variation of ResNet backbones for breast mass detection alongside different weight initialization, mainly ImageNet, COCO weights, and model trained from scratch. ...

Transfer learning in breast mass detection and classification

Journal of Ambient Intelligence and Humanized Computing

... In particular, algorithmic developments have been proposed for the detection of micro-dumps in satellite VHR images and to automatically generate unvalidated map products for subsequent validation by photointerpreters ( Fig. 1). In particular, both traditional machine learning [4] and deep learning [5] based techniques have been recently proposed and validated. Compared to aerial orthophoto [6], the usage of satellite detection products faces considerable challenges due to the limited spatial resolution and accuracy, resulting in a scarcity of literature on this subject. ...

Illegal Microdumps Detection in Multi-Mission Satellite Images With Deep Neural Network and Transfer Learning Approach

IEEE Access

... The results showed that the overall accuracy of detection was 96.33%, and the overall accuracy of classification was 85.52%. Kassahun and his team [8] designed a system for the detection, segmentation, and classification of breast lumps, in which the detection phase used the YOLO model for the initial detection of breast lumps. In the publicly available RadImageNet dataset, using DenseNet-121 combined with the YOLOv5m model, the IoU threshold was 0.5, and the mAP was 0.718. ...

Breast Mass Detection and Classification Using Transfer Learning on OPTIMAM Dataset Through RadImageNet Weights
  • Citing Chapter
  • January 2024

Lecture Notes in Computer Science

... An application based on ML models can be used for analyzing complex datasets to discover the patterns in diseases that exceed the level of intuitive judgments of human doctors (34). For example, in the diagnosis of lung cancer (35,36), breast cancer (37)(38)(39), and many other types of cancers (40), the ML models have been experimentally proven to be equal to or even better than a radiologist in the diagnosis in some cases. A ML system can automate data analysis and, thus, the time between detection to diagnosis is shorter, especially for the diseases which need a rapid response like stroke and heart attack (41). ...

Editorial: AI applications for diagnosis of breast cancer

Frontiers in Artificial Intelligence

... Existing automated methods remain fundamentally constrained, exhibiting significant vulnerabilities to noise interference and signal discontinuities. Consequently, these approaches frequently necessitate extensive manual intervention and post-processing to rectify reconstruction artifacts and validate morphological accuracy [29,30]. ...

BigNeuron: a resource to benchmark and predict performance of algorithms for automated tracing of neurons in light microscopy datasets

Nature Methods

... As observed in the table, research primarily focuses on mass detection [13,17,20,22], mass classification (benign or malignant) [14,26,29,30], mass segmentation [18], and anomaly detection [32], often utilizing either open or private datasets. In terms of architecture selection, there is a notable trend moving from two-stage object detection methods [14][15][16]18]toward single-stage [27,[29][30][31] or transformerbased [28] approaches. Single-stage models, in particular, are favored for their faster prediction performance and easier implementation. ...

Transformer-based mass detection in digital mammograms

Journal of Ambient Intelligence and Humanized Computing

... ViT has demonstrated remarkable performance in image classification, even surpassing traditional architectures like ResNets [31]. Inspired by ViT, researchers have developed variations such as the swin transformer, which adapts the ResNet-50 architecture to create a hierarchical ViT [32]. These adaptations aim to enhance the original ViT design by integrating more recent training techniques without introducing additional attentionbased modules. ...

Convolutional Networks and Transformers for Mammography Classification: An Experimental Study

... In the 1990s, image partitioning was synonymous with the watershed transform, which has been utilized for segmentation and feature extraction [3,4,5,6]. The watershed has also been used as a pre-or post-processing step in various deep learning (DL) applications [7,8,9,10,11]. However, watersheds have two main drawbacks: over-segmentation (regions are too small) and slow processing times (if executed sequentially). ...

Seamless Iterative Semi-supervised Correction of Imperfect Labels in Microscopy Images
  • Citing Chapter
  • September 2022

Lecture Notes in Computer Science