Hellena Hempe’s research while affiliated with University of Lübeck and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (8)


To determine the most suitable architecture for our task, we employ combinations of several encoder–decoder architectures including traditional convolutional methods and geometric methods. The AEs are trained to reconstruct either a point cloud representation or a volumetric surface representation of vertebrae, which are derived from the previously computed segmentation mask. As Shape Encoder (A), we employ a convolutional method, as well as a point-based and a graph-based method to predict the embedding z. As Shape Decoder (B), we employ a convolutional method as well as a point-based method and propose a novel point-based shape decoder. The Shape Classifier (C) is then trained separately on the embedding z for each encoder–decoder combination using the same multilayer perceptron (MLP) model. Note that only the weights of the MLP are trained in a supervised manner, whereas the weights of the encoder are fixed.
Point-based shape decoder: From the embedding vector z, a point representation of N key points is computed using an MLP. The layers each consist of a 1D convolution with the channel size denoted by white font within the blocks, InstanceNorm and ReLU. The number on top of the blocks denotes the size of the dimensionality of the point cloud. Afterwards, a differentiable sampling operation is applied on the key points to obtain a volumetric representation. This step requires N additional parameters y.
ROC curve and corresponding AUC for encoder–decoder combinations of the median AUC of 10 seeds. The encoders are grouped by color and line style, whereas the decoders are grouped by color and marker. The corresponding area under curve (AUC) is listed inside the legend.
Results of our data-hold-out experiment as boxplots and scatterplots of the AUC obtained for 10 random seeds each. The plots are separated by the employed encoder architecture, and provide the classification results obtained with the respective decoder. Top-Left: Convolutional encoder Top-Right: Point-encoder models, Bottom-Left: Graph-encoder models Bottom-Right: End-to-End trained models including the traditional CNN trained on image intensities instead and trained on vertebra surface (denoted as Conv-encoder img and surf).
Assessment of architectural building blocks employed within our framework includ- ing the number of trainable parameters, total size of the model and computational demands in Mult-Add operations.
Shape Matters: Detecting Vertebral Fractures Using Differentiable Point-Based Shape Decoding
  • Article
  • Full-text available

February 2024

·

24 Reads

·

3 Citations

Hellena Hempe

·

Alexander Bigalke

·

Mattias Paul Heinrich

Background: Degenerative spinal pathologies are highly prevalent among the elderly population. Timely diagnosis of osteoporotic fractures and other degenerative deformities enables proactive measures to mitigate the risk of severe back pain and disability. Methods: We explore the use of shape auto-encoders for vertebrae, advancing the state of the art through robust automatic segmentation models trained without fracture labels and recent geometric deep learning techniques. Our shape auto-encoders are pre-trained on a large set of vertebrae surface patches. This pre-training step addresses the label scarcity problem faced when learning the shape information of vertebrae for fracture detection from image intensities directly. We further propose a novel shape decoder architecture: the point-based shape decoder. Results: Employing segmentation masks that were generated using the TotalSegmentator, our proposed method achieves an AUC of 0.901 on the VerSe19 testset. This outperforms image-based and surface-based end-to-end trained models. Our results demonstrate that pre-training the models in an unsupervised manner enhances geometric methods like PointNet and DGCNN. Conclusion: Our findings emphasize the advantages of explicitly learning shape features for diagnosing osteoporotic vertebrae fractures. This approach improves the reliability of classification results and reduces the need for annotated labels.

Download

DeepSTAPLE: Learning to Predict Multimodal Registration Quality for Unsupervised Domain Adaptation

July 2022

·

23 Reads

Lecture Notes in Computer Science

·

Alexander Bigalke

·

Christian N. Kruse

·

[...]

·

While deep neural networks often achieve outstanding results on semantic segmentation tasks within a dataset domain, performance can drop significantly when predicting domain-shifted input data. Multi-atlas segmentation utilizes multiple available sample annotations which are deformed and propagated to the target domain via multimodal image registration and fused to a consensus label afterwards but subsequent network training with the registered data may not yield optimal results due to registration errors. In this work, we propose to extend a curriculum learning approach with additional regularization and fixed weighting to train a semantic segmentation model along with data parameters representing the atlas confidence. Using these adjustments we can show that registration quality information can be extracted out of a semantic segmentation model and further be used to create label consensi when using a straightforward weighting scheme. Comparing our results to the STAPLE method, we find that our consensi are not only a better approximation of the oracle-label regarding Dice score but also improve subsequent network training results. KeywordsDomain adaptationMulti-atlas registrationLabel noiseConsensusCurriculum learning




Overview of the Heidelberg Colorectal (HeiCo) data set. Raw data comprises anonymized, downsampled laparoscopic video data from three different types of colorectal surgery along with corresponding streams from medical devices in the operating room. Annotations include surgical phase information for the entire video sequences as well as information on instrument presence and corresponding instance-wise segmentation masks of medical instruments (if any) for more than 10,000 frames.
Laparoscopic images representing various levels of difficulty for the tasks of medical instrument detection, binary segmentation and multi-instance segmentation. Raw input frames (a) and corresponding reference segmentation masks (b) computed from the reference contours.
Examples of challenging frames overlaid with reference multi-instance segmentations created by surgical data science experts.
Folder structure for the complete data set. It comprises five levels corresponding to (1) surgery type, (2) procedure number, (3) procedural data (video and device data along with phase annotations), (4) frame number and (5) frame-based data.
Folder structure for the ROBUST-MIS challenge data set. It comprises five levels corresponding to (1) data type (training/test), (2) surgery type, (3) procedure number, (4) frame number and (5) case data.
Heidelberg colorectal data set for surgical data science in the sensor operating room

April 2021

·

1,041 Reads

·

109 Citations

Scientific Data

Image-based tracking of medical instruments is an integral part of surgical data science applications. Previous research has addressed the tasks of detecting, segmenting and tracking medical instruments based on laparoscopic video data. However, the proposed methods still tend to fail when applied to challenging images and do not generalize well to data they have not been trained on. This paper introduces the Heidelberg Colorectal (HeiCo) data set - the first publicly available data set enabling comprehensive benchmarking of medical instrument detection and segmentation algorithms with a specific emphasis on method robustness and generalization capabilities. Our data set comprises 30 laparoscopic videos and corresponding sensor data from medical devices in the operating room for three different types of laparoscopic surgery. Annotations include surgical phase labels for all video frames as well as information on instrument presence and corresponding instance-wise segmentation masks for surgical instruments (if any) in more than 10,000 individual frames. The data has successfully been used to organize international competitions within the Endoscopic Vision Challenges 2017 and 2019.


Comparative validation of multi-instance instrument segmentation in endoscopy: Results of the ROBUST-MIS 2019 challenge

November 2020

·

249 Reads

·

88 Citations

Medical Image Analysis

Intraoperative tracking of laparoscopic instruments is often a prerequisite for computer and robotic-assisted interventions. While numerous methods for detecting, segmenting and tracking of medical instruments based on endoscopic video images have been proposed in the literature, key limitations remain to be addressed: Firstly, robustness, that is, the reliable performance of state-of-the-art methods when run on challenging images (e.g. in the presence of blood, smoke or motion artifacts). Secondly, generalization; algorithms trained for a specific intervention in a specific hospital should generalize to other interventions or institutions. In an effort to promote solutions for these limitations, we organized the Robust Medical Instrument Segmentation (ROBUST-MIS) challenge as an international benchmarking competition with a specific focus on the robustness and generalization capabilities of algorithms. For the first time in the field of endoscopic image processing, our challenge included a task on binary segmentation and also addressed multi-instance detection and segmentation. The challenge was based on a surgical data set comprising 10,040 annotated images acquired from a total of 30 surgical procedures from three different types of surgery. The validation of the competing methods for the three tasks (binary segmentation, multi-instance detection and multi-instance segmentation) was performed in three different stages with an increasing domain gap between the training and the test data. The results confirm the initial hypothesis, namely that algorithm performance degrades with an increasing domain gap. While the average detection and segmentation quality of the best-performing algorithms is high, future research should concentrate on detection and segmentation of small, crossing, moving and transparent instrument(s) (parts).


Fig. 3: Examples of challenging frames overlaid with reference multi-instance segmentations created by surgical data science experts.
Fig. 4: Folder structure for the complete data set. It comprises five levels corresponding to (1) surgery type, (2) procedure number, (3) procedural data (video and device data along with phase annotations) and (4) frame number and (5) frame-based data.
Medical devices of the operating room and corresponding sensor streams provided by the OR1 FUSION® (KARL STORZ SE & Co KG, Tuttlingen, Germany).
Number of frames selected from the different procedures.
Heidelberg Colorectal Data Set for Surgical Data Science in the Sensor Operating Room

May 2020

·

402 Reads

Image-based tracking of medical instruments is an integral part of many surgical data science applications. Previous research has addressed the tasks of detecting, segmenting and tracking medical instruments based on laparoscopic video data. However, the methods proposed still tend to fail when applied to challenging images and do not generalize well to data they have not been trained on. This paper introduces the Heidelberg Colorectal (HeiCo) data set - the first publicly available data set enabling comprehensive benchmarking of medical instrument detection and segmentation algorithms with a specific emphasis on robustness and generalization capabilities of the methods. Our data set comprises 30 laparoscopic videos and corresponding sensor data from medical devices in the operating room for three different types of laparoscopic surgery. Annotations include surgical phase labels for all frames in the videos as well as instance-wise segmentation masks for surgical instruments in more than 10,000 individual frames. The data has successfully been used to organize international competitions in the scope of the Endoscopic Vision Challenges (EndoVis) 2017 and 2019.


Robust Medical Instrument Segmentation Challenge 2019

March 2020

·

891 Reads

·

4 Citations

Intraoperative tracking of laparoscopic instruments is often a prerequisite for computer and robotic-assisted interventions. While numerous methods for detecting, segmenting and tracking of medical instruments based on endoscopic video images have been proposed in the literature, key limitations remain to be addressed: Firstly, robustness, that is, the reliable performance of state-of-the-art methods when run on challenging images (e.g. in the presence of blood, smoke or motion artifacts). Secondly, generalization; algorithms trained for a specific intervention in a specific hospital should generalize to other interventions or institutions. In an effort to promote solutions for these limitations, we organized the Robust Medical Instrument Segmentation (ROBUST-MIS) challenge as an international benchmarking competition with a specific focus on the robustness and generalization capabilities of algorithms. For the first time in the field of endoscopic image processing, our challenge included a task on binary segmentation and also addressed multi-instance detection and segmentation. The challenge was based on a surgical data set comprising 10,040 annotated images acquired from a total of 30 surgical procedures from three different types of surgery. The validation of the competing methods for the three tasks (binary segmentation, multi-instance detection and multi-instance segmentation) was performed in three different stages with an increasing domain gap between the training and the test data. The results confirm the initial hypothesis, namely that algorithm performance degrades with an increasing domain gap. While the average detection and segmentation quality of the best-performing algorithms is high, future research should concentrate on detection and segmentation of small, crossing, moving and transparent instrument(s) (parts).

Citations (5)


... Radi-ologists use the Genant scale (Genant et al., 1993) to measure fracture severity from CT images. Deep Learning can automate VCF detection (Valentinitsch et al., 2019;Tomita et al., 2018;Chettrit et al., 2020;Husseini et al., 2020b;Yilmaz et al., 2021;Engstler et al., 2022;Windsor et al., 2022;Iyer et al., 2023;Hempe et al., 2024), however, only a few works have considered VCF grading, all fully-supervised (Pisov et al., 2020;Zakharov et al., 2023;Wei et al., 2022;Yilmaz et al., 2023). Compared to fracture detection, grading is an even more imbalanced task since medium to severely fractured vertebrae account for only a small portion of overall data. ...

Reference:

Counterfactual Explanations for Medical Image Classification and Regression using Diffusion Autoencoder
Shape Matters: Detecting Vertebral Fractures Using Differentiable Point-Based Shape Decoding

... However, significant advancements have been made in recent years with the advent of deep learning methods, particularly convolutional neural networks, which have revolutionized spine medical image segmentation. 22,23 Models such as U-Net, 24 DeepLab, 25 and Fully Convolutional Networks 26 have been extensively applied to spine image segmentation tasks, with the U-Net architecture achieving notable success. This model excels in extracting and representing feature information in MRI medical image analysis. ...

Opportunistic CT screening for degenerative deformities and osteoporotic fractures with 3D DeepLab
  • Citing Conference Paper
  • April 2022

... We created a new surgical VQA benchmarking dataset for basic visual perception tasks by applying our framework to the existing Heidelberg Colorectal (HeiCo) dataset [14]. HeiCo comprises 30 laparoscopic surgical videos (10 each from proctocolectomy, rectal resection, and sigmoid resection procedures) with 10,040 frames containing instance segmentations. ...

Heidelberg colorectal data set for surgical data science in the sensor operating room

Scientific Data

... Previous attempts by Nakamura 35 and Tokuyasu 36 , using the YoloV3 algorithm for multi-object detection during gallbladder resection without relying on pixel-level segmentation resulted in accuracy ranging only from 0.07 to 0.32 35 segmentation. The multiclass segmentation of the SMV-PV axis and splenic vein obtained a mean Dice coefficient of around 0.5, indicating the challenges due to: 1) The indistinct visual contrast between those anatomical landmarks contributed to lower recognition accuracy, whereas the mean Dice coefficient for identification of gallbladder triangle is notably higher; 2) Obstructions of tissues or instruments often segmented the boundary of anatomic landmark structures, dividing them into several smaller segments, thereby complicating the recognition; 3) Simultaneous objects in multiclass semantic segmentation tasks may lead to a significant decrease in recognition accuracy 37 . Therefore, we merged the anatomical landmarks and applied binary segmentation for better model performance. ...

Comparative validation of multi-instance instrument segmentation in endoscopy: Results of the ROBUST-MIS 2019 challenge

Medical Image Analysis

... Detecting and tracking surgical instruments in laparoscopic videos is crucial for autonomous surgery and enhanced clinical support [1]. The trend in the field is toward the utilization of deep learning methodologies [2,3] 1 models heavily depend on fully supervised learning, requiring extensive annotated data. ...

Robust Medical Instrument Segmentation Challenge 2019