Marcel Bengs’s research while affiliated with Hamburg University of Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (46)


Nodule Detection and Generation on Chest X-Rays: NODE21 Challenge
  • Article

March 2024

·

29 Reads

·

12 Citations

IEEE Transactions on Medical Imaging

·

Bram Van Ginneken

·

·

[...]

·

Keelin Murphy

Pulmonary nodules may be an early manifestation of lung cancer, the leading cause of cancer-related deaths among both men and women. Numerous studies have established that deep learning methods can yield high-performance levels in the detection of lung nodules in chest X-rays. However, the lack of gold-standard public datasets slows down the progression of the research and prevents benchmarking of methods for this task. To address this, we organized a public research challenge, NODE21, aimed at the detection and generation of lung nodules in chest X-rays. While the detection track assesses state-of-the-art nodule detection systems, the generation track determines the utility of nodule generation algorithms to augment training data and hence improve the performance of the detection systems. This paper summarizes the results of the NODE21 challenge and performs extensive additional experiments to examine the impact of the synthetically generated nodule training images on the detection algorithm performance.


Schematic visualization of our deep learning pipeline for nodule detection. Left: Nodule generation process. A nodule is embedded in a nodule-free chest x-ray at a given position and scale. Middle: Training of multiple model architectures, independently of each other. Right: Evaluation and ensembling of the trained models. The predictions of each model are merged via weighted box fusion (WBF)⁶² which results in one aggregated prediction for all models.
Analysis of bounding boxes in the data. From left to right: Position of all bounding boxes on the x and y axis height plotted against the width of all bounding boxes, respectively, exemplary x-ray without nodules, exemplary x-ray with one nodule.
Left: Correctly predicted nodule. Middle and Right: Correctly predicted nodule together with false positive predictions. Yellow arrows indicate the groundtruth location of nodules.
Left: AUROC of all individual models and their ensemble, evaluated on Dadd\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\mathscr {D}_{add}$$\end{document}. Right: FROC of all individual models and their ensemble, capped at an average of two false positives per patient.
A systematic approach to deep learning-based nodule detection in chest radiographs
  • Article
  • Full-text available

June 2023

·

129 Reads

·

14 Citations

Lung cancer is a serious disease responsible for millions of deaths every year. Early stages of lung cancer can be manifested in pulmonary lung nodules. To assist radiologists in reducing the number of overseen nodules and to increase the detection accuracy in general, automatic detection algorithms have been proposed. Particularly, deep learning methods are promising. However, obtaining clinically relevant results remains challenging. While a variety of approaches have been proposed for general purpose object detection, these are typically evaluated on benchmark data sets. Achieving competitive performance for specific real-world problems like lung nodule detection typically requires careful analysis of the problem at hand and the selection and tuning of suitable deep learning models. We present a systematic comparison of state-of-the-art object detection algorithms for the task of lung nodule detection. In this regard, we address the critical aspect of class imbalance and and demonstrate a data augmentation approach as well as transfer learning to boost performance. We illustrate how this analysis and a combination of multiple architectures results in state-of-the-art performance for lung nodule detection, which is demonstrated by the proposed model winning the detection track of the Node21 competition. The code for our approach is available at https://github.com/FinnBehrendt/node21-submit.

Download


Real-Time Motion Analysis With 4D Deep Learning for Ultrasound-Guided Radiotherapy

March 2023

·

35 Reads

·

7 Citations

IEEE transactions on bio-medical engineering

Motion compensation in radiation therapy is a challenging scenario that requires estimating and forecasting motion of tissue structures to deliver the target dose. Ultrasound offers direct imaging of tissue in real-time and is considered for image guidance in radiation therapy. Recently, fast volumetric ultrasound has gained traction, but motion analysis with such high-dimensional data remains difficult. While deep learning could bring many advantages, such as fast data processing and high performance, it remains unclear how to process sequences of hundreds of image volumes efficiently and effectively. We present a 4D deep learning approach for real-time motion estimation and forecasting using long-term 4D ultrasound data. Using motion traces acquired during radiation therapy combined with various tissue types, our results demonstrate that long-term motion estimation can be performed markerless with a tracking error of 0.35±0.20.35\pm 0.2 mm and with an inference time of less than 5 ms. Also, we demonstrate forecasting directly from the image data up to 900 ms into the future. Overall, our findings highlight that 4D deep learning is a promising approach for motion analysis during radiotherapy.


Optical force estimation for interactions between tool and soft tissues

January 2023

·

94 Reads

·

12 Citations

Robotic assistance in minimally invasive surgery offers numerous advantages for both patient and surgeon. However, the lack of force feedback in robotic surgery is a major limitation, and accurately estimating tool-tissue interaction forces remains a challenge. Image-based force estimation offers a promising solution without the need to integrate sensors into surgical tools. In this indirect approach, interaction forces are derived from the observed deformation, with learning-based methods improving accuracy and real-time capability. However, the relationship between deformation and force is determined by the stiffness of the tissue. Consequently, both deformation and local tissue properties must be observed for an approach applicable to heterogeneous tissue. In this work, we use optical coherence tomography, which can combine the detection of tissue deformation with shear wave elastography in a single modality. We present a multi-input deep learning network for processing of local elasticity estimates and volumetric image data. Our results demonstrate that accounting for elastic properties is critical for accurate image-based force estimation across different tissue types and properties. Joint processing of local elasticity information yields the best performance throughout our phantom study. Furthermore, we test our approach on soft tissue samples that were not present during training and show that generalization to other tissue properties is possible.


Supervised Contrastive Learning to Classify Paranasal Anomalies in the Maxillary Sinus

September 2022

·

34 Reads

·

7 Citations

Lecture Notes in Computer Science

Using deep learning techniques, anomalies in the paranasal sinus system can be detected automatically in MRI images and can be further analyzed and classified based on their volume, shape and other parameters like local contrast. However due to limited training data, traditional supervised learning methods often fail to generalize. Existing deep learning methods in paranasal anomaly classification have been used to diagnose at most one anomaly. In our work, we consider three anomalies. Specifically, we employ a 3D CNN to separate maxillary sinus volumes without anomaly from maxillary sinus volumes with anomaly. To learn robust representations from a small labelled dataset, we propose a novel learning paradigm that combines contrastive loss and cross-entropy loss. Particularly, we use a supervised contrastive loss that encourages embeddings of maxillary sinus volumes with and without anomaly to form two distinct clusters while the cross-entropy loss encourages the 3D CNN to maintain its discriminative ability. We report that optimising with both losses is advantageous over optimising with only one loss. We also find that our training strategy leads to label efficiency. With our method, a 3D CNN classifier achieves an AUROC of 0.85 ± 0.03 while a 3D CNN classifier optimised with cross-entropy loss achieves an AUROC of 0.66 ± 0.1. Our source code is available at https://github.com/dawnofthedebayan/SupConCE_MICCAI_22.


Supervised Contrastive Learning to Classify Paranasal Anomalies in the Maxillary Sinus

September 2022

·

68 Reads

Using deep learning techniques, anomalies in the paranasal sinus system can be detected automatically in MRI images and can be further analyzed and classified based on their volume, shape and other parameters like local contrast. However due to limited training data, traditional supervised learning methods often fail to generalize. Existing deep learning methods in paranasal anomaly classification have been used to diagnose at most one anomaly. In our work, we consider three anomalies. Specifically, we employ a 3D CNN to separate maxillary sinus volumes without anomalies from maxillary sinus volumes with anomalies. To learn robust representations from a small labelled dataset, we propose a novel learning paradigm that combines contrastive loss and cross-entropy loss. Particularly, we use a supervised contrastive loss that encourages embeddings of maxillary sinus volumes with and without anomaly to form two distinct clusters while the cross-entropy loss encourages the 3D CNN to maintain its discriminative ability. We report that optimising with both losses is advantageous over optimising with only one loss. We also find that our training strategy leads to label efficiency. With our method, a 3D CNN classifier achieves an AUROC of 0.85 while a 3D CNN classifier optimised with cross-entropy loss achieves an AUROC of 0.66.


Averaged performance metrics. The l1-error is evaluated on the healthy test sets. The DICE and AUPRC values are provided in percent.
Capturing Inter-Slice Dependencies of 3D Brain MRI-Scans for Unsupervised Anomaly Detection

May 2022

·

158 Reads

·

8 Citations

The increasing workloads for radiologists in clinical practice lead to the need for an automatic support tool for anomaly detection in brain MRI-scans. While supervised learning methods can detect and localize lesions in brain MRI-scans, the need for large, balanced data sets with pixel-level annotations limits their use. In contrast, unsupervised anomaly detection (UAD) models only require healthy brain data for training. Despite the inherent 3D structure of brain MRI-scans, most UAD studies focus on slice-wise processing. In this work, we capture the inter-slice dependencies of the human brain using recurrent neural networks (RNN) and transformer-based self-attention mechanisms together with variational autoencoders (VAE). We show that by this we can improve both reconstruction quality and UAD performance while the number of parameters remain similar to the 2D approach where the slices are processed individually.


Fig. 3 Exemplary trajectory for patient liver motion during free breathing in radiotherapy. The motion for x (blue), y (yellow) and z (red) is shown (a) as well as the main motion component of the three dimensions after applying a PCA (b)
Fig. 4 Slices from US volumes showing the spherical marker (a and b) and exemplary bovine tissue (c and d) with 8 × 8 beams and 16 × 16 beams, respectively. The red boxes indicate the crop used for tracking
Fig. 5 Results for NCC (red, dotted) and MOSSE (green, dotted) for step-and-shoot data set with the ground truth (blue). The main motion component of the trajectories and tracking results are shown for trajectory 4 for 16 × 16 beams (a) and 8 × 8 beams (b)
Mean and standard deviation of the amplitude, as well as min- imum and maximum amplitude from the main motion component are reported in mm as well as the number of breathing cycles and the dura- tion of the measurements in seconds of the trajectories for the markerless liver measurements
Tracking error e t in mm for evaluation of the different US data with MOSSE and NCC for the spherical marker. Continuously acquired data is compared to data acquired in a step-and-shoot fashion, as well as the different number of beams used for imaging
Systematic analysis of volumetric ultrasound parameters for markerless 4D motion tracking

May 2022

·

49 Reads

·

2 Citations

International Journal of Computer Assisted Radiology and Surgery

Objectives Motion compensation is an interesting approach to improve treatments of moving structures. For example, target motion can substantially affect dose delivery in radiation therapy, where methods to detect and mitigate the motion are widely used. Recent advances in fast, volumetric ultrasound have rekindled the interest in ultrasound for motion tracking. We present a setup to evaluate ultrasound based motion tracking and we study the effect of imaging rate and motion artifacts on its performance. Methods We describe an experimental setup to acquire markerless 4D ultrasound data with precise ground truth from a robot and evaluate different real-world trajectories and system settings toward accurate motion estimation. We analyze motion artifacts in continuously acquired data by comparing to data recorded in a step-and-shoot fashion. Furthermore, we investigate the trade-off between the imaging frequency and resolution. Results The mean tracking errors show that continuously acquired data leads to similar results as data acquired in a step-and-shoot fashion. We report mean tracking errors up to 2.01 mm and 1.36 mm on the continuous data for the lower and higher resolution, respectively, while step-and-shoot data leads to mean tracking errors of 2.52 mm and 0.98 mm. Conclusions We perform a quantitative analysis of different system settings for motion tracking with 4D ultrasound. We can show that precise tracking is feasible and additional motion in continuously acquired data does not impair the tracking. Moreover, the analysis of the frequency resolution trade-off shows that a high imaging resolution is beneficial in ultrasound tracking.


Fig. 5. Our setup for ultrasound shear wave data acquisition. (Left) A linear ultrasound probe is positioned by the robot on the black gelatin block. (Right) The different push locations relative to the ROI.
Fig. 8. Full image elasticity predictions. Shown are the standard deviations of the pixelwise estimated Young's moduli of 40 push and image sequences. Results for ground truth elasticity 37.55 kPa and 72.64kPa are shown in the top and bottom row, respectively. The black square indicates the ROI used for training, the push location is indicated by the black dashed line. We use a spatio-temporal window size of 65 × 65 × 35.
Fig. 9. Elasticity maps of five inclusion shapes. Column 1 shows the target Young's modulus, column 2-4 spatio-temporal CNN predictions with spatio-temporal window sizes of 17 × 17 pixels, 33 × 33 pixels and 65 × 65 pixels. Time-of-flight estimates are depicted in column 5.
Ultrasound Shear Wave Elasticity Imaging With Spatio-Temporal Deep Learning

April 2022

·

73 Reads

·

14 Citations

IEEE transactions on bio-medical engineering

Ultrasound shear wave elasticity imaging is a valuable tool for quantifying the elastic properties of tissue. Typically, the shear wave velocity is derived and mapped to an elasticity value, which neglects information such as the shape of the propagating shear wave or push sequence characteristics. We present 3D spatio-temporal CNNs for fast local elasticity estimation from ultrasound data. This approach is based on retrieving elastic properties from shear wave propagation within small local regions. A large training data set is acquired with a robot from homogeneous gelatin phantoms ranging from 17.42 kPa to 126.05 kPa with various push locations. The results show that our approach can estimate elastic properties on a pixelwise basis with a mean absolute error of 5.014.37 kPa. Furthermore, we estimate local elasticity independent of the push location and can even perform accurate estimates inside the push region. For phantoms with embedded inclusions, we report a 53.93% lower MAE (7.50 kPa) and on the background of 85.24% (1.64 kPa) compared to a conventional shear wave method. Overall, our method offers fast local estimations of elastic properties with small spatio-temporal window sizes.


Citations (27)


... ms, TR=4.05 ms, matrix size = 260x320, in-plane resolution=1.09 mm × 1.09 mm and slice thickness=3 mm.Mask-RCNN 212 has been used for various types of tumor detection and segmentation40,48,220 . We employed this framework for the current task of liver GTV detection + segmentation with reconstructed accelerated images. ...

Reference:

Artificial Intelligence Augmented Medical Imaging Reconstruction in Radiation Therapy
Nodule Detection and Generation on Chest X-Rays: NODE21 Challenge
  • Citing Article
  • March 2024

IEEE Transactions on Medical Imaging

... Pulmonary nodules refer to circular or irregular lesions with a diameter less than or equal to 3cm in the lung, which can be divided into benign and malignant types [6,7]. Chest X-ray imaging is considered to be one of the effective means to detect pulmonary nodules [8]. Although existing methods have made remarkable achievements in the classification, detection and segmentation of single tasks, many models are still limited to processing a single task and lack the ability to simultaneously perform multi-task prediction and comprehensive analysis of multiple state indicators of nodules. ...

A systematic approach to deep learning-based nodule detection in chest radiographs

... With advancements in the field of computer vision, artificial intelligence is expected to improve real-time, non-invasive target tracking, dose reconstruction, and treatment plan adjustment [15]. Markerless tracking, which utilizes deep learning utilizing various types of real-time medical images such as ultrasound, 4D CT, and X-ray, has been reported in numerous studies [16][17][18]. Additionally, accurate and rapid motion prediction algorithms can be used to deal with the delays caused by system positioning and action time. ...

Real-Time Motion Analysis With 4D Deep Learning for Ultrasound-Guided Radiotherapy
  • Citing Article
  • March 2023

IEEE transactions on bio-medical engineering

... Accurate intraoperative OCE could then give physicians back the ability to feel for changes in elastic properties during minimally invasive procedures. In addition to localizing pathological tissue, knowledge of biomechanical properties also enables better monitoring of tool-tissue interactions, e.g. in vision-based force estimation [62]. Finally, any OCT system inherently captures temporal sequences of A-scans. ...

Optical force estimation for interactions between tool and soft tissues

... Convolutional neural networks (CNNs) are recognized for diagnosing paranasal pathologies, evidenced in sinusitis classification [10,11], differentiating inverted papilloma from carcinomas [12], and detecting MS fungal ball and chronic rhinosinusitis in CT scans [13]. Prior studies have explored contrastive learning and cross-entropy loss for MS anomaly classification [14], and MS extraction techniques from MRI [15]. However, all of the aforementioned methods use supervised learning. ...

Supervised Contrastive Learning to Classify Paranasal Anomalies in the Maxillary Sinus
  • Citing Chapter
  • September 2022

Lecture Notes in Computer Science

... Unsupervised anomaly detection (UAD), which involves modeling the distribution of normal data and identifying deviations as anomalies, has gained attention as a promising alternative [7,11,35,42]. Conventional unsupervised methods, such as autoencoders [32] and generative adversarial networks (GANs) [14], attempt to reconstruct normal anatomical structures and flag areas with high reconstruction errors as anomalies. Despite their potential, these approaches suffer from notable limitations. ...

Unsupervised Anomaly Detection in 3D Brain MRI Using Deep Learning with Impured Training Data
  • Citing Conference Paper
  • March 2022

... We evaluate the effectiveness of our method against Thresh [21], AE [9], VAE [9], SVAE [22], DAE [7], f-AnoGAN [13], DDPM [8], mDDPM [15] and pDDPM [14], in terms of Dice- In Table I, we compare our pDDPM-IQA with state-of-theart methods on BraTS21 and MSLUB using T2 modality in a cross-dataset setting, as adopted in previous studies [14], [15]. Our pDDPM-IQA significantly (p < 0.05) outperforms all baseline approaches on both datasets in terms of DICE and AUPRC, with improvements exceeding 10%. ...

Capturing Inter-Slice Dependencies of 3D Brain MRI-Scans for Unsupervised Anomaly Detection

... Researchers at Stanford and Duke University have addressed the increasing demand for financially accessible alternatives and have developed setup devices that do not involve the high costs of the 3D ultrasound equipment mentioned above, but further improvements are needed [7,9]. Current studies are focused on increasing the accuracy and intuitiveness of ultrasound image acquisition, with less reliance on operator skills; this way, several emerging tracking technologies -although still in the research stage, have proved significant potential to overcome current difficulties regarding motion artifacts and imaging resolution [11,12]. In the era of the artificial intelligence breakthrough, it has become increasingly necessary to reach high rates of reproducibility in the field of diagnostic imaging by promoting automatization as part of medical progress. ...

Systematic analysis of volumetric ultrasound parameters for markerless 4D motion tracking

International Journal of Computer Assisted Radiology and Surgery

... Lastly, as with many areas of healthcare, there are opportunities for artificial intelligence and machine or deep learning techniques to aid in improving SWE technology [173][174][175]. Several groups have investigated methods that use machine learning for classification based on SWE data, use deep learning methods to reconstruct shear wave velocity maps, and estimate viscoelastic parameters based on shear wave motion signals. ...

Ultrasound Shear Wave Elasticity Imaging With Spatio-Temporal Deep Learning

IEEE transactions on bio-medical engineering

... Bengs et al. (2021) compared 3D and 2D VAEs for anomaly detection on brain MRI. Bengs et al. (2022) trained a VAE on 3D T1-weighted MRI by additionally considering the age information. Simarro Viana et al. (2020) proposed a 3D extension of the 2D f-AnoGAN and refined the training steps to detect traumatic brain injuries. ...

Unsupervised anomaly detection in 3D brain MRI using deep learning with multi-task brain age prediction
  • Citing Conference Paper
  • April 2022