TMI202106论⽂汇总(IEEETransactionsonMedicalImaging)1. Segmentation-Renormalized Deep Feature Modulation for Unpaired Image Harmonization
Mengwei Ren. Neel Dey. James Fishbaugh. Guido Gerig.
激流三部曲“
Deep networks are now ubiquitous in large-scale multi-center imaging studies. However, the direct aggregation of images across sites is contraindicated for downstream statistical and deep learning-based image analysis due to inconsistent contrast, resolution, and noise. To this end, in the absence of paired data, variations of Cycle-consistent Generative Adversarial Networks have been used to harmonize image sets between a source and target domain.贺卡设计
Importantly, these methods are prone to instability, contrast inversion, intractable manipulation of pathology, and steganographic mappings which limit their reliable adoption in real-world medical imaging. In this work, based on an underlying assumption that morphological shape is consistent across imaging sites, we propose a segmentation-renormalized image translation framework to reduce inter-scanner heterogeneity while preserving anatomical layout.
We replace the affine transformations used in the normalization layers within generative networks with trainable scale and shift parameters conditioned on jointly learned anatomical segmentation embeddings to modulate features at every level of translation. We evaluate our methodologies against recent baselines across several imaging modalities (T1w MRI, FLAIR MRI, and OCT) on datasets with and without lesions. Segmentation-renormalization for translation GANs yields superior image harmonization as quantified by Inception distances, demonstrates improved downstream utility via post-hoc segmentation accuracy, and improved robustness to translation perturbation and self-adversarial attacks.
2. A Cervical Histopathology Dataset for Computer Aided Diagnosis of Precancerous Lesions
⽤于计算机辅助诊断癌前病变的宫颈组织病理学数据集
Zhu Meng. Zhicheng Zhao. Bingyang Li. Fei Su. Limei Guo.
“
Cervical cancer, as one of the most frequently diagnosed cancers worldwide, is curable when detected early.
Histopathology images play an important role in precision medicine of the cervical lesions. However, few computer aided algorithms have been explored on cervical histopathology images due to the lack of public datasets. In this article, we release a new cervical histopathology image dataset for automated precancerous diagnosis. Specifically, 100 slides from 71 patients are annotated by three independent pathologists. To show the difficulty of the task, benchmarks are obtained through both fully and weakly supervised learning. Extensive experiments based on typical classification and semantic segmentation networks are carried out to provide strong baselines. In particular, a strategy of assembling classification, segmentation, and pseudo-labeling is proposed to further improve the performance. The Dice coefficient reaches 0.7833, indicating the feasibility of computer aided diagnosis and the effectiveness of our weakly supervised ensemble algorithm. The dataset and evaluation codes are publicly available. To the best of our knowledge, it is the first public cervical histopathology dataset for automated precancerous segmentation. We believe that this work will attract researchers to explore novel algorithms on cervical automated diagnosis, thereby assisting doctors and patients clinically.
3. Learning With Context Feedback Loop for Robust Medical Image Segmentation
使⽤上下⽂反馈循环进⾏稳健医学图像分割的学习
Kibrom Berihu Girum. Gilles Créhange. Alain Lalande.
Deep learning has successfully been leveraged for medical image segmentation. It employs convolutional neural networks (CNN) to learn distinctive image features from a defined pixel-wise objective function. However, this
approach can lead to less output pixel interdependence producing incomplete and unrealistic segmentation results. In this paper, we present a fully automatic deep learning method for robust medical image segmentation by formulating the segmentation problem as a recurrent framework using two systems. The first one is a forward system of an encoder-decoder CNN that predicts the segmentation result from the input image. The predicted probabilistic output of the forward system is then encoded by a fully convolutional network (FCN)-based context feedback system. The encoded feature space of the FCN is then integrated back into the forward system’s feed-forward learning process.
Using the FCN-based context feedback loop allows the forward system to learn and extract more high-level image features and fix previous mistakes, thereby improving prediction accuracy over time. Experimental results, performed on four different clinical datasets, demonstrate our method’s p
otential application for single and multi-structure medical image segmentation by outperforming the state of the art methods. With the feedback loop, deep learning methods can now produce results that are both anatomically plausible and robust to low contrast images. Therefore, formulating image segmentation as a recurrent framework of two interconnected networks via context feedback loop can be a potential method for robust and efficient medical image analysis.
4. Cascaded Regression Neural Nets for Kidney Localization and Segmentation-free Volume Estimation
⽤于肾脏定位和⽆分割体积估计的级联回归神经⽹络
Mohammad Arafat Hussain. Ghassan Hamarneh. Rafeef Garbi.
“
Kidney volume is an essential biomarker for a number of kidney disease diagnoses, for example, chronic kidney disease. Existing total kidney volume estimation methods often rely on an intermediate kidney segmentation step. On the other hand, automatic kidney localization in volumetric medical images is a critical step that often precedes subsequent data processing and ana
lysis. Most current approaches perform kidney localization via an intermediate classification or regression step. This paper proposes an integrated deep learning approach for (i) kidney localization in computed tomography scans and (ii) segmentation-free renal volume estimation. Our localization method uses a selection -convolutional neural network that approximates the kidney inferior-superior span along the axial direction.
Cross-sectional (2D) slices from the estimated span are subsequently used in a combined sagittal-axial Mask-RCNN that detects the organ bounding boxes on the axial and sagittal slices, the combination of which produces a final 3D organ bounding box. Furthermore, we use a fully convolutional network to estimate the kidney volume that skips the segmentation procedure. We also present a mathematical expression to approximate the ‘volume error’ metric from the ‘Sørensen–Dice coefficient.’ We accessed 100 patients’ CT scans from the Vancouver General Hospital records and obtained 210 patients’ CT scans from the 2019 Kidney Tumor Segmentation Challenge database to validate our method. Our method produces a kidney boundary wall localization error of ~2.4mm and a mean volume estimation error of ~5%.
5. Direct Differentiation of Pathological Changes in the Human Lung Parenchyma With Grating-Based Spectral X-ray Dark-Field Radiography
⽤基于光栅的光谱 X 射线暗场射线照相直接鉴别⼈肺实质的病理变化
Kirsten Taphorn. Korbinian Mechlem. Thorsten Sellerer. Fabio De Marco. Manuel Viermetz. Franz Pfeiffer. Daniela Pfeiffer. Julia Herzen.
Diagnostic lung imaging is often associated with high radiation dose and lacks sensitivity, especially for diagnosing early stages of structural lung diseases. Therefore, diagnostic imaging methods are required which provide sound diagnosis of lung diseases with a high sensitivity as well as low patient dose. In small animal experiments, the
sensitivity of grating-based X-ray dark-field imaging to structural changes in the lung tissue was demonstrated. The energy-dependence of the X-ray dark-field signal of lung tissue is a function of its microstructure and not yet known.
Furthermore, conventional X-ray dark-field imaging is not capable of differentiating different types of pathological changes, such as fibrosis and emphysema. Here we demonstrate the potential diagnostic power of grating-based X-ray dark-field in combination with spectral imaging in human chest radiography for the direct differentiation of lung diseases. We investigated the energy-dependent linear diffusion coefficient of simulated lung tissue with different diseases in wave-propag
ation simulations and validated the results with analytical calculations. Additionally, we modeled spectral X-ray dark-field chest radiography scans to exploit these differences in energy-dependency. The results demonstrate the potential to directly differentiate structural changes in the human lung. Consequently, grating-based spectral X-ray dark-field imaging potentially contributes to the differential diagnosis of structural lung diseases at a clinically relevant dose level.
6. Development and Initial Results of a Brain PET Insert for Simultaneous 7-Tesla PET/MRI Using an FPGA-Only Signal Digitization Method
使⽤仅 FPGA 的信号数字化⽅法开发⽤于同时进⾏ 7-Tesla PET/MRI 的脑 PET 插⼊物的开发和初步结果
静夜星空Jun Yeon Won. Haewook Park. Seungeun Lee. Jeong-Whan Son. Yina Chung. Guen Bae Ko. Kyeong Yun Kim. Junghyun Song. Seongho Seo. Yeunchul Ryu. Jun-Young Chung. Jae Sung Lee.月经期可以染发吗
“
In study, we developed a positron emission tomography (PET) insert for simultaneous brain imaging within 7-Tesla (7T) magnetic resonance (MR) imaging scanners. The PET insert has 18 sectors, and
each sector is assembled with two-layer depth-of-interaction (DOI)-capable high-resolution block detectors. The PET scanner features a 16.7-cm-long axial field-of-view (FOV) to provide entire human brain images without bed movement. The PET scanner early digitizes
a large number of block detector signals at a front-end data acquisition (DAQ) board using a novel field-programmable
gate array (FPGA)-only signal digitization method. All the digitized PET data from the front-end DAQ boards are transferred using gigabit transceivers via non-magnetic high-definition multimedia interface (HDMI) cables. A back-end DAQ system provides a common clock and synchronization signal for FPGAs over the HDMI cables. An active cooling system using copper heat pipes is applied for thermal regulation. All the 2.17-mm-pitch crystals with two-layer DOI information were clearly identified in the block detectors, exhibiting a system-level energy resolution of 12.6%. The PET scanner yielded clear hot-rod and Hoffman brain phantom images and demonstrated 3D PET imaging capability without bed movement. We also performed a pilot simultaneous PET/MR imaging study of a brain phantom. The PET scanner achieved a spatial resolution of 2.5 mm at the center FOV (NU 4) and a sensitivity of 18.9 kcps/MBq (NU 2) and 6.19% (NU 4) in accordance with the National Electrical Manufacturers Association (NEMA) standards.
7. Multi-Modal Retinal Image Classification With Modality-Specific Attention Network
具有特定模态注意⽹络的多模态视⽹膜图像分类
Xingxin He. Ying Deng. Leyuan Fang. Qinghua Peng.
Recently, automatic diagnostic approaches have been widely used to classify ocular diseases. Most of these
approaches are based on a single imaging modality (e.g., fundus photography or optical coherence tomography (OCT)), which usually only reflect the oculopathy to a certain extent, and neglect the modality-specific information among different imaging modalities. This paper proposes a novel modality-specific attention network (MSAN) for multi-modal retinal image classification, which can effectively utilize the modality-specific diagnostic features from fundus and OCT images. The MSAN comprises two attention modules to extract the modality-specific features from fundus and OCT images, respectively. Specifically, for the fundus image, ophthalmologists need to observe local and global pathologies at multiple scales (e.g., from microaneurysms at the micrometer level, optic disc at millimeter level to blood vessels through the whole eye). Therefore, we propose a multi-scale attention module to extract both the local and global features from fundus images. Moreover, large b
ackground regions exist in the OCT image, which is meaningless for diagnosis. Thus, a region-guided attention module is proposed to encode the retinal layer-related features and ignore the background in OCT images. Finally, we fuse the modality-specific features to form a multi-modal feature and train the multi-modal retinal image classification network. The fusion of modality-specific features allows the model to combine the advantages of fundus and OCT modality for a more accurate diagnosis. Experimental results on
a clinically acquired multi-modal retinal image (fundus and OCT) dataset demonstrate that our MSAN outperforms
other well-known single-modal and multi-modal retinal image classification methods.
8. Learning Tubule-Sensitive CNNs for Pulmonary Airway and Artery-Vein Segmentation in CT
在 CT 中学习⽤于肺⽓道和动脉静脉分割的⼩管敏感 CNN
曲率中心Yulei Qin. Hao Zheng. Yun Gu. Xiaolin Huang. Jie Yang. Lihui Wang. Feng Yao. Yue-Min Zhu. Guang-Zhong Yang.
“
9. 3D Multi-Attention Guided Multi-Task Learning Network for Automatic Gastric Tumor Segmentation and Lymph Node Classification
⽤于⾃动胃肿瘤分割和淋巴结分类的 3D 多注意⼒引导多任务学习⽹络
Yongtao Zhang. Haimei Li. Jie Du. Jing Qin. Tianfu Wang. Yue Chen. Bing Liu. Wenwen Gao. Guolin Ma. Baiying Lei.
“
10. Relation-Induced Multi-Modal Shared Representation Learning for Alzheimer’s Disease Diagnosis
⽤于阿尔茨海默病诊断的关系诱导的多模态共享表⽰学习
Zhenyuan Ning. Qing Xiao. Qianjin Feng. Wufan Chen. Yu Zhang.
The fusion of multi-modal data (e.g., magnetic resonance imaging (MRI) and positron emission tomography (PET)) has been prevalent for accurate identification of Alzheimer’s disease (AD) by providing complementary structural and functional information. However, most of the existing method
s simply concatenate multi-modal features in the original space and ignore their underlying associations which may provide more discriminative characteristics for AD
identification. Meanwhile, how to overcome the overfitting issue caused by high-dimensional multi-modal data remains appealing. To this end, we propose a relation-induced multi-modal shared representation learning method for AD diagnosis. The proposed method integrates representation learning, dimension reduction, and classifier modeling into a unified framework. Specifically, the framework first obtains multi-modal shared representations by learning a bi-directional mapping between original space and shared space. Within this shared space, we utilize several relational regularizers (including feature-feature, feature-label, and sample-sample regularizers) and auxiliary regularizers to encourage learning underlying associations inherent in multi-modal data and alleviate overfitting, respectively. Next, we project the shared representations into the target space for AD diagnosis. To validate the effectiveness of our proposed approach, we conduct extensive experiments on two independent datasets (i.e., ADNI-1 and ADNI-2), and the experimental results demonstrate that our proposed method outperforms several state-of-the-art methods.
11. Hierarchical Temporal Attention Network for Thyroid Nodule Recognition Using Dynamic CEUS Imaging
红豆泡多久使⽤动态 CEUS 成像进⾏甲状腺结节识别的分层时间注意⽹络
Peng Wan. Fang Chen. Chunrui Liu. Wentao Kong. Daoqiang Zhang.
“
Contrast-enhanced ultrasound (CEUS) has emerged as a popular imaging modality in thyroid nodule diagnosis due to its ability to visualize vascular distribution in real time. Recently, a number of learning-based methods are dedicated to mine pathological-related enhancement dynamics and make prediction at one step, ignoring a native diagnostic dependency. In clinics, the differentiation of benign or malignant nodules always precedes the recognition of
pathological types. In this paper, we propose a novel hierarchical temporal attention network (HiTAN) for thyroid nodule diagnosis using dynamic CEUS imaging, which unifies dynamic enhancement feature learning and hierarchical nodules classification into a deep framework. Specifically, this method decomposes the diagnosis of nodules into an ordered two-stage classification task, where diagnostic dependency is modeled by Gated Recurrent Units (GRUs).
Besides, we design a local-to-global temporal aggregation (LGTA) operator to perform a comprehen
sive temporal fusion along the hierarchical prediction path. Particularly, local temporal information is defined as typical enhancement patterns identified with the guidance of perfusion representation learned from the differentiation level. Then, we leverage an attention mechanism to embed global enhancement dynamics into each identified salient pattern. In this study, we evaluate the proposed HiTAN method on the collected CEUS dataset of thyroid nodules. Extensive
experimental results validate the efficacy of dynamic patterns learning, fusion and hierarchical diagnosis mechanism.
12. Measurement of the Lag Correction Factor in Low-Dose Fluoroscopic Imaging
低剂量透视成像中滞后校正因⼦的测量
Dong Sik Kim. Eunae Lee.
发布评论