Dorin Comaniciu’s research while affiliated with New Jersey Institute of Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (471)


Figure 1. Workflow of using a biomarker-based pre-screening test.
Figure 4. Comorbidity analysis on (a) colorectal, (b) liver, and (c) lung cancer validation cohorts. Likelihood ratio plots show the likelihood ratios of corresponding models on subgroups of patients suffering from the indicated comorbidities, for increasing risk thresholds on the testing dataset. Barplots illustrate the significance of association of the corresponding comorbidity with cancer type (as negative logarithm of the p-value obtained with Fisher's exact test) over the entire dataset. The prevalence of the respective comorbidities in the cancer-positive and control cases is listed with each comorbidity. Note that the base prevalence is different for each comorbidity. GI, gastrointestinal.
Deep learning-based identification of patients at increased risk of cancer using routine laboratory markers
  • Preprint
  • File available

October 2024

·

8 Reads

·

Shikha Chaganti

·

Matthias Siebert

·

[...]

·

Early screening for cancer has proven to improve the survival rate and spare patients from intensive and costly treatments due to late diagnosis. Cancer screening in the healthy population involves an initial risk stratification step to determine the screening method and frequency, primarily to optimize resource allocation by targeting screening towards individuals who draw most benefit. For most screening programs, age and clinical risk factors such as family history are part of the initial risk stratification algorithm. In this paper, we focus on developing a blood marker-based risk stratification approach, which could be used to identify patients with elevated cancer risk to be encouraged for taking a diagnostic test or participate in a screening program. We demonstrate that the combination of simple, widely available blood tests, such as complete blood count and complete metabolic panel, could potentially be used to identify patients at risk for colorectal, liver, and lung cancers with areas under the ROC curve of 0.76, 0.85, 0.78, respectively. Furthermore, we hypothesize that such an approach could not only be used as pre-screening risk assessment for individuals but also as population health management tool, for example to better interrogate the cancer risk in certain sub-populations.

Download


Fig. 2. Downstream tasks. Cine view classification: a) SAX, b) 2CH, c) 3CH, d) 4CH, e-f) Aorta, g-h) Other, i) Cine SAX segmentation, j) Cine LAX 4CH seg-mentation, k-l) LGE SAX segmentation, m-o) Mapping segmentation (Pre-contrast T1, postcontrast T1, T2, respectively) , p) Anterior and Inferior RVIP localization on cine SAX, q-u) Disease detection (NORM, DCM, HCM, abnormal RV, MINF respectively), v-x) LGE detection (No LGE, LGE present, LGE present, respectively). Abbreviations: SAX: short axis; LAX: long axis, CH: chamber, RVIP: RV insertion point; NORM: Normal; DCM: Dilated Cardiomyopathy; HCM: Hypertrophic cardiomyopathy; MINF: myocardial infarction.
Towards a vision foundation model for comprehensive assessment of Cardiac MRI

October 2024

·

19 Reads

Cardiac magnetic resonance imaging (CMR), considered the gold standard for noninvasive cardiac assessment, is a diverse and complex modality requiring a wide variety of image processing tasks for comprehensive assessment of cardiac morphology and function. Advances in deep learning have enabled the development of state-of-the-art (SoTA) models for these tasks. However, model training is challenging due to data and label scarcity, especially in the less common imaging sequences. Moreover, each model is often trained for a specific task, with no connection between related tasks. In this work, we introduce a vision foundation model trained for CMR assessment, that is trained in a self-supervised fashion on 36 million CMR images. We then finetune the model in supervised way for 9 clinical tasks typical to a CMR workflow, across classification, segmentation, landmark localization, and pathology detection. We demonstrate improved accuracy and robustness across all tasks, over a range of available labeled dataset sizes. We also demonstrate improved few-shot learning with fewer labeled samples, a common challenge in medical image analyses. We achieve an out-of-box performance comparable to SoTA for most clinical tasks. The proposed method thus presents a resource-efficient, unified framework for CMR assessment, with the potential to accelerate the development of deep learning-based solutions for image analysis tasks, even with few annotated data available.


Deep Learning-based Unsupervised Domain Adaptation via a Unified Model for Prostate Lesion Detection Using Multisite Biparametric MRI Datasets

August 2024

·

11 Reads

Radiology Artificial Intelligence

Purpose To determine whether the unsupervised domain adaptation (UDA) method with generated images improves the performance of a supervised learning (SL) model for prostate cancer (PCa) detection using multisite biparametric (bp) MRI datasets. Materials and Methods This retrospective study included data from 5150 patients (14 191 samples) collected across nine different imaging centers. A novel UDA method using a unified generative model was developed for PCa detection using multisite bpMRI datasets. This method translates diffusion-weighted imaging (DWI) acquisitions, including apparent diffusion coefficient (ADC) and individual diffusion-weighted (DW) images acquired using various b values, to align with the style of images acquired using b values recommended by Prostate Imaging Reporting and Data System (PI-RADS) guidelines. The generated ADC and DW images replace the original images for PCa detection. An independent set of 1692 test cases (2393 samples) was used for evaluation. The area under the receiver operating characteristic curve (AUC) was used as the primary metric, and statistical analysis was performed via bootstrapping. Results For all test cases, the AUC values for baseline SL and UDA methods were 0.73 and 0.79 (P < .001), respectively, for PCa lesions with PI-RADS score of 3 or greater and 0.77 and 0.80 (P < .001) for lesions with PI-RADS scores of 4 or greater. In the 361 test cases under the most unfavorable image acquisition setting, the AUC values for baseline SL and UDA were 0.49 and 0.76 (P < .001) for lesions with PI-RADS scores of 3 or greater and 0.50 and 0.77 (P < .001) for lesions with PI-RADS scores of 4 or greater. Conclusion UDA with generated images improved the performance of SL methods in PCa lesion detection across multisite datasets with various b values, especially for images acquired with significant deviations from the PI-RADS-recommended DWI protocol (eg, with an extremely high b value). Keywords: Prostate Cancer Detection, Multisite, Unsupervised Domain Adaptation, Diffusion-weighted Imaging, b Value Supplemental material is available for this article. © RSNA, 2024.


Deep Learning-based Unsupervised Domain Adaptation via a Unified Model for Prostate Lesion Detection Using Multisite Bi-parametric MRI Datasets

August 2024

·

34 Reads

Our hypothesis is that UDA using diffusion-weighted images, generated with a unified model, offers a promising and reliable strategy for enhancing the performance of supervised learning models in multi-site prostate lesion detection, especially when various b-values are present. This retrospective study included data from 5,150 patients (14,191 samples) collected across nine different imaging centers. A novel UDA method using a unified generative model was developed for multi-site PCa detection. This method translates diffusion-weighted imaging (DWI) acquisitions, including apparent diffusion coefficient (ADC) and individual DW images acquired using various b-values, to align with the style of images acquired using b-values recommended by Prostate Imaging Reporting and Data System (PI-RADS) guidelines. The generated ADC and DW images replace the original images for PCa detection. An independent set of 1,692 test cases (2,393 samples) was used for evaluation. The area under the receiver operating characteristic curve (AUC) was used as the primary metric, and statistical analysis was performed via bootstrapping. For all test cases, the AUC values for baseline SL and UDA methods were 0.73 and 0.79 (p<.001), respectively, for PI-RADS>=3, and 0.77 and 0.80 (p<.001) for PI-RADS>=4 PCa lesions. In the 361 test cases under the most unfavorable image acquisition setting, the AUC values for baseline SL and UDA were 0.49 and 0.76 (p<.001) for PI-RADS>=3, and 0.50 and 0.77 (p<.001) for PI-RADS>=4 PCa lesions. The results indicate the proposed UDA with generated images improved the performance of SL methods in multi-site PCa lesion detection across datasets with various b values, especially for images acquired with significant deviations from the PI-RADS recommended DWI protocol (e.g. with an extremely high b-value).




Multi-Agent Reinforcement Learning Meets Leaf Sequencing in Radiotherapy

June 2024

·

4 Reads

In contemporary radiotherapy planning (RTP), a key module leaf sequencing is predominantly addressed by optimization-based approaches. In this paper, we propose a novel deep reinforcement learning (DRL) model termed as Reinforced Leaf Sequencer (RLS) in a multi-agent framework for leaf sequencing. The RLS model offers improvements to time-consuming iterative optimization steps via large-scale training and can control movement patterns through the design of reward mechanisms. We have conducted experiments on four datasets with four metrics and compared our model with a leading optimization sequencer. Our findings reveal that the proposed RLS model can achieve reduced fluence reconstruction errors, and potential faster convergence when integrated in an optimization planner. Additionally, RLS has shown promising results in a full artificial intelligence RTP pipeline. We hope this pioneer multi-agent RL leaf sequencer can foster future research on machine learning for RTP.


Self-supervised learning for interventional image analytics: toward robust device trackers

May 2024

·

6 Reads

·

1 Citation

Journal of Medical Imaging

Purpose: The accurate detection and tracking of devices, such as guiding catheters in live X-ray image acquisitions, are essential prerequisites for endovascular cardiac interventions. This information is leveraged for procedural guidance, e.g., directing stent placements. To ensure procedural safety and efficacy, there is a need for high robustness/no failures during tracking. To achieve this, one needs to efficiently tackle challenges, such as device obscuration by the contrast agent or other external devices or wires and changes in the field-of-view or acquisition angle, as well as the continuous movement due to cardiac and respiratory motion. Approach: To overcome the aforementioned challenges, we propose an approach to learn spatio-temporal features from a very large data cohort of over 16 million interventional X-ray frames using self-supervision for image sequence data. Our approach is based on a masked image modeling technique that leverages frame interpolation-based reconstruction to learn fine inter-frame temporal correspondences. The features encoded in the resulting model are fine-tuned downstream in a light-weight model. Results: Our approach achieves state-of-the-art performance, in particular for robustness, compared to ultra optimized reference solutions (that use multi-stage feature fusion or multi-task and flow regularization). The experiments show that our method achieves a 66.31% reduction in the maximum tracking error against the reference solutions (23.20% when flow regularization is used), achieving a success score of 97.95% at a 3× faster inference speed of 42 frames-per-second (on GPU). In addition, we achieve a 20% reduction in the standard deviation of errors, which indicates a much more stable tracking performance. Conclusions: The proposed data-driven approach achieves superior performance, particularly in robustness and speed compared with the frequently used multi-modular approaches for device tracking. The results encourage the use of our approach in various other tasks within interventional image analytics that require effective understanding of spatio-temporal semantics.


Computation of proposed positional embedding (PPE).
The mechanism of physical-coordinate-based positional embedding.
Vision transformer architecture with co-ordinate-based positional embedding. Proposed positional embedding is added to each modality separately followed by the creation of non-overlapping patches. Linear projection is used on the flattened patches and patch-wise positional embedding is added before sending these patches to the transformer encoder.
Qualitative comparison of different methods with and without proposed positional embedding. An illustrative segmentation example of the predicted labels which demonstrate differences in methods. First row consists of single slice of AxTrace, AxADC and corresponding infarct ground truth. Second row consists of segmentation output of different methods.
A typical segmentation example of the predicted labels using proposed co-ordinate based embedding. The first row depicts around 75th percentile performance of two samples based on the Dice score. Second and third rows depict around 50 percentile and around 25 percentile performance respectively.
Co-ordinate-based positional embedding that captures resolution to enhance transformer’s performance in medical image analysis

April 2024

·

23 Reads

Vision transformers (ViTs) have revolutionized computer vision by employing self-attention instead of convolutional neural networks and demonstrated success due to their ability to capture global dependencies and remove spatial biases of locality. In medical imaging, where input data may differ in size and resolution, existing architectures require resampling or resizing during pre-processing, leading to potential spatial resolution loss and information degradation. This study proposes a co-ordinate-based embedding that encodes the geometry of medical images, capturing physical co-ordinate and resolution information without the need for resampling or resizing. The effectiveness of the proposed embedding is demonstrated through experiments with UNETR and SwinUNETR models for infarct segmentation on MRI dataset with AxTrace and AxADC contrasts. The dataset consists of 1142 training, 133 validation and 143 test subjects. Both models with the addition of co-ordinate based positional embedding achieved substantial improvements in mean Dice score by 6.5% and 7.6%. The proposed embedding showcased a statistically significant advantage p-value< 0.0001 over alternative approaches. In conclusion, the proposed co-ordinate-based pixel-wise positional embedding method offers a promising solution for Transformer-based models in medical image analysis. It effectively leverages physical co-ordinate information to enhance performance without compromising spatial resolution and provides a foundation for future advancements in positional embedding techniques for medical applications.


Citations (61)


... However, designing a clinically acceptable treatment plan is still a complicated and extremely time-consuming process, which is highly dependent on the experience of medical physicists [3], [4]. One solution for this limitation is to automatically predict the dose distribution of radiotherapy by extracting knowledge from existing plans, thereby providing the standard for dosimetry verification and quality control of future radiotherapy plans [5][6][7]. ...

Reference:

ARANet: Attention-based Residual Adversarial Network with Deep Supervision for Radiotherapy Dose Prediction of Cervical Cancer
Flexible-C m GAN: Towards Precise 3D Dose Prediction in Radiotherapy
  • Citing Conference Paper
  • June 2023

... memory module 10 . Manuela et al. 18 proposed an approach for radiology report generation that is a two-step method that primarily detects abnormalities in chest X-ray (CXR) images. This initial step addresses a multiclass problem by localizing identified abnormalities with bounding boxes and associated probability scores and detecting various lung lesions such as nodules, masses, and pneumothorax in X-rays 18 . ...

Generation of Radiology Findings in Chest X-Ray by Leveraging Collaborative Knowledge
  • Citing Article
  • January 2023

Procedia Computer Science

... Besides, MIS-FM [137] introduced a pretext task based on pseudo-segmentation, where Volume Fusion (VF) was proposed to generate paired images and segmentation labels to pre-train the 3D segmentation model. Ghesu et al. [138] proposed a method for self-supervised learning based on CL and online feature clustering. d) HL-based pre-training combine various pre-training approaches to fuse their advantages in joint training. ...

Contrastive self-supervised learning from 100 million medical images with optional supervision
  • Citing Article
  • November 2022

Journal of Medical Imaging

... "It broke the division of space and the discipline of power on the body by constructing a squarelike public dialogue space, and provided an appropriate scene for people to express their views, communicate their opinions and participate in politics." [2] All users from different cultures and classes participate in the "Square" to show their personalities and construct new meanings and scenes in the discussion. ...

A Validation Study of the Deep-Learning-Based Prostate Imaging Reporting and Data System Scoring Algorithm

Open Journal of Radiology

... An increasing number of AI-based models have been developed for the diagnosis and prognosis of COVID-19 using both CT scans and radiographs. These models utilize AI-based methods to extract COVID-19 relevant features, such as lung patterns and disease-specific manifestations [51,52,[55][56][57]. A retrospective study conducted in a tertiary hospital in the United States has reported that AI-read chest radiographs could potentially have comparable prognostic performance to CT scans [55]. ...

Value of quantitative airspace disease measured on chest CT and chest radiography at initial diagnosis compared to clinical variables for prediction of severe COVID-19
  • Citing Article
  • June 2022

Journal of Medical Imaging

... 3 Especially the detection of ICH on brain CT has very promising results. [4][5][6] However, clinical implementation of these systems remains scarce, and there is need for substantial external validation in real-world settings in order to ensure sufficient performance in novel clinical environments. 7,8 In fact, in a recent systematic review of external validation studies of deep learning algorithms for image-based radiological diagnosis, 9 the vast majority of these algorithms demonstrated diminished diagnostic performance on external datasets, with some reporting a substantial performance decrease. ...

AI with Statistical Confidence Scores for Detection of Acute/Subacute Hemorrhage in Noncontrast Head CT Scans
  • Citing Article
  • April 2022

Radiology Artificial Intelligence

... Attempts have been made to construct such foundation models in the field of radiology. One group of researchers trained a vision foundation model on 100 million medical images, including radiographs, CT images, MR images, and ultrasound images [67]. Another group of researchers trained a self-supervised network on 4.8 million chest radiographs [68]. ...

Self-supervised Learning from 100 Million Medical Images
  • Citing Preprint
  • January 2022

... Thus, ML tools can learn to predict outcomes based on a known severity metric or establish a new severity scale that improves the known risk assessment methods by incorporating novel features not included in the standardized scales [17]. Furthermore, past studies have focused on predicting the severity conditions of in-patients, such as MV requirements [18][19][20][21], days spent in ICU [20][21][22][23][24], whether rehospitalization is necessary for recurring health problems [15,25], and forecasting the need for other therapeutic interventions [24]. ...

A Deep Learning Approach for Predicting Severity of COVID-19 Patients Using A Parsimonious Set of Laboratory Markers

iScience

... [3] Another method involves measuring the maximum distance from the midline formed by the anterior falx and posterior falx to the septum pellucidum at the axial level of the foramen of Monro. [3,6,8] Another approach suggested measuring the distance from a line connecting the most anterior and posterior visible points on the falx to the farthest point on the septum pellucidum. [9] In recent years, there has been a growing interest in leveraging artificial intelligence (AI) techniques for automating radiological procedures in TBI cases. ...

Brain midline shift detection and quantification by a cascaded deep network pipeline on non-contrast computed tomography scans
  • Citing Conference Paper
  • October 2021

... A characteristic symptom of severe ICH is brain midline shift (MLS), which is the lateral displacement of midline cerebral structures (see Fig. 1). MLS is an important and quantifiable indicator of the severity of mass effects and the urgency of intervention [2,3,9]. For instance, the 5 millimeters (mm) threshold of MLS is frequently used to determine whether immediate intervention and close monitoring is required [4]. ...

Brain midline shift detection and quantification by a cascaded deep network pipeline on non-contrast computed tomography scans
  • Citing Article
  • November 2021