15 research outputs found
Capsules for Biomedical Image Segmentation
Our work expands the use of capsule networks to the task of object
segmentation for the first time in the literature. This is made possible via
the introduction of locally-constrained routing and transformation matrix
sharing, which reduces the parameter/memory burden and allows for the
segmentation of objects at large resolutions. To compensate for the loss of
global information in constraining the routing, we propose the concept of
"deconvolutional" capsules to create a deep encoder-decoder style network,
called SegCaps. We extend the masked reconstruction regularization to the task
of segmentation and perform thorough ablation experiments on each component of
our method. The proposed convolutional-deconvolutional capsule network,
SegCaps, shows state-of-the-art results while using a fraction of the
parameters of popular segmentation networks. To validate our proposed method,
we perform experiments segmenting pathological lungs from clinical and
pre-clinical thoracic computed tomography (CT) scans and segmenting muscle and
adipose (fat) tissue from magnetic resonance imaging (MRI) scans of human
subjects' thighs. Notably, our experiments in lung segmentation represent the
largest-scale study in pathological lung segmentation in the literature, where
we conduct experiments across five extremely challenging datasets, containing
both clinical and pre-clinical subjects, and nearly 2000 computed-tomography
scans. Our newly developed segmentation platform outperforms other methods
across all datasets while utilizing less than 5% of the parameters in the
popular U-Net for biomedical image segmentation. Further, we demonstrate
capsules' ability to generalize to unseen rotations/reflections on natural
images.Comment: Extension of the non-archival Capsules of Object Segmentation with
experiments on both clinical and pre-clinical pathological lung segmentation
from CT scans and muscular and adipose tissue segmentation from MR images.
Accepted for publication in Medical Image Analysis. DOI:
https://doi.org/10.1016/j.media.2020.101889. arXiv admin note: text overlap
with arXiv:1804.0424
Deep Learning for Musculoskeletal Image Analysis
The diagnosis, prognosis, and treatment of patients with musculoskeletal
(MSK) disorders require radiology imaging (using computed tomography, magnetic
resonance imaging(MRI), and ultrasound) and their precise analysis by expert
radiologists. Radiology scans can also help assessment of metabolic health,
aging, and diabetes. This study presents how machinelearning, specifically deep
learning methods, can be used for rapidand accurate image analysis of MRI
scans, an unmet clinicalneed in MSK radiology. As a challenging example, we
focus on automatic analysis of knee images from MRI scans and study machine
learning classification of various abnormalities including meniscus and
anterior cruciate ligament tears. Using widely used convolutional neural
network (CNN) based architectures, we comparatively evaluated the knee
abnormality classification performances of different neural network
architectures under limited imaging data regime and compared single and
multi-view imaging when classifying the abnormalities. Promising results
indicated the potential use of multi-view deep learning based classification of
MSK abnormalities in routine clinical assessment.Comment: Invited Paper, ASILOMAR 2019, TP4b: Machine Learning Advances in
Computational Imagin
Neural Transformers for Intraductal Papillary Mucosal Neoplasms (IPMN) Classification in MRI images
Early detection of precancerous cysts or neoplasms, i.e., Intraductal
Papillary Mucosal Neoplasms (IPMN), in pancreas is a challenging and complex
task, and it may lead to a more favourable outcome. Once detected, grading
IPMNs accurately is also necessary, since low-risk IPMNs can be under
surveillance program, while high-risk IPMNs have to be surgically resected
before they turn into cancer. Current standards (Fukuoka and others) for IPMN
classification show significant intra- and inter-operator variability, beside
being error-prone, making a proper diagnosis unreliable. The established
progress in artificial intelligence, through the deep learning paradigm, may
provide a key tool for an effective support to medical decision for pancreatic
cancer. In this work, we follow this trend, by proposing a novel AI-based IPMN
classifier that leverages the recent success of transformer networks in
generalizing across a wide variety of tasks, including vision ones. We
specifically show that our transformer-based model exploits pre-training better
than standard convolutional neural networks, thus supporting the sought
architectural universalism of transformers in vision, including the medical
image domain and it allows for a better interpretation of the obtained results
Towards Automatic Cartilage Quantification in Clinical Trials - Continuing from the 2019 IWOAI Knee Segmentation Challenge.
OBJECTIVE: To evaluate whether the deep learning (DL) segmentation methods from the six teams that participated in the IWOAI 2019 Knee Cartilage Segmentation Challenge are appropriate for quantifying cartilage loss in longitudinal clinical trials. DESIGN: We included 556 subjects from the Osteoarthritis Initiative study with manually read cartilage volume scores for the baseline and 1-year visits. The teams used their methods originally trained for the IWOAI 2019 challenge to segment the 1130 knee MRIs. These scans were anonymized and the teams were blinded to any subject or visit identifiers. Two teams also submitted updated methods. The resulting 9,040 segmentations are available online.The segmentations included tibial, femoral, and patellar compartments. In post-processing, we extracted medial and lateral tibial compartments and geometrically defined central medial and lateral femoral sub-compartments. The primary study outcome was the sensitivity to measure cartilage loss as defined by the standardized response mean (SRM). RESULTS: For the tibial compartments, several of the DL segmentation methods had SRMs similar to the gold standard manual method. The highest DL SRM was for the lateral tibial compartment at 0.38 (the gold standard had 0.34). For the femoral compartments, the gold standard had higher SRMs than the automatic methods at 0.31/0.30 for medial/lateral compartments. CONCLUSION: The lower SRMs for the DL methods in the femoral compartments at 0.2 were possibly due to the simple sub-compartment extraction done during post-processing. The study demonstrated that state-of-the-art DL segmentation methods may be used in standardized longitudinal single-scanner clinical trials for well-defined cartilage compartments
Capsules for biomedical image segmentation
Xu, Ziyue/0000-0002-5728-6869; IRMAKCI, ISMAIL/0000-0003-3277-2710WOS:000613293400001PubMed: 33246227Our work expands the use of capsule networks to the task of object segmentation for the first time in the literature. This is made possible via the introduction of locally-constrained routing and transformation matrix sharing, which reduces the parameter/memory burden and allows for the segmentation of objects at large resolutions. To compensate for the loss of global information in constraining the routing, we propose the concept of "deconvolutional" capsules to create a deep encoder-decoder style network, called SegCaps. We extend the masked reconstruction regularization to the task of segmentation and perform thorough ablation experiments on each component of our method. The proposed convolutional-deconvolutional capsule network, SegCaps, shows state-of-the-art results while using a fraction of the parameters of popular segmentation networks. To validate our proposed method, we perform experiments segmenting pathological lungs from clinical and pre-clinical thoracic computed tomography (CT) scans and segmenting muscle and adipose (fat) tissue from magnetic resonance imaging (MRI) scans of human subjects' thighs. Notably, our experiments in lung segmentation represent the largest-scale study in pathological lung segmentation in the literature, where we conduct experiments across five extremely challenging datasets, containing both clinical and pre-clinical subjects, and nearly 2000 computed-tomography scans. Our newly developed segmentation platform outperforms other methods across all datasets while utilizing less than 5% of the parameters in the popular U-Net for biomedical image segmentation. Further, we demonstrate capsules' ability to generalize to unseen handling of rotations/reflections on natural images. Published by Elsevier B.V.NIHUnited States Department of Health & Human ServicesNational Institutes of Health (NIH) - USA [R01-EB020539, R01-CA246704]This study is partially supported by the NIH grant R01-EB020539 and R01-CA246704
Multi-Contrast MRI Segmentation Trained on Synthetic Images
© 2022 IEEE.In our comprehensive experiments and evaluations, we show that it is possible to generate multiple contrast (even all synthetically) and use synthetically generated images to train an image segmentation engine. We showed promising segmentation results tested on real multi-contrast MRI scans when delineating muscle, fat, bone and bone marrow, all trained on synthetic images. Based on synthetic image training, our segmentation results were as high as 93.91%, 94.11%, 91.63%, 95.33%, for muscle, fat, bone, and bone marrow delineation, respectively. Results were not significantly different from the ones obtained when real images were used for segmentation training: 94.68%, 94.67%, 95.91%, and 96.82%, respectively. Clinical relevance - Synthetically generated images could potentially be used in large-scale training of deep networks for segmentation purpose. Small data set problem of many clinical imaging problems can potentially be addressed with the proposed algorithm
