Title | Siamese Encoder-based Spatial-Temporal Mixer for Growth Trend Prediction of Lung Nodules on CT Scans |
Author | |
Corresponding Author | Liu,Jiang |
DOI | |
Publication Years | 2022
|
Conference Name | 25th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI)
|
ISSN | 0302-9743
|
EISSN | 1611-3349
|
ISBN | 978-3-031-16430-9
|
Source Title | |
Volume | 13431 LNCS
|
Pages | 484-494
|
Conference Date | SEP 18-22, 2022
|
Conference Place | null,Singapore,SINGAPORE
|
Publication Place | GEWERBESTRASSE 11, CHAM, CH-6330, SWITZERLAND
|
Publisher | |
Abstract | In the management of lung nodules, we are desirable to predict nodule evolution in terms of its diameter variation on Computed Tomography (CT) scans and then provide a follow-up recommendation according to the predicted result of the growing trend of the nodule. In order to improve the performance of growth trend prediction for lung nodules, it is vital to compare the changes of the same nodule in consecutive CT scans. Motivated by this, we screened out 4,666 subjects with more than two consecutive CT scans from the National Lung Screening Trial (NLST) dataset to organize a temporal dataset called NLSTt. In specific, we first detect and pair regions of interest (ROIs) covering the same nodule based on registered CT scans. After that, we predict the texture category and diameter size of the nodules through models. Last, we annotate the evolution class of each nodule according to its changes in diameter. Based on the built NLSTt dataset, we propose a siamese encoder to simultaneously exploit the discriminative features of 3D ROIs detected from consecutive CT scans. Then we novelly design a spatial-temporal mixer (STM) to leverage the interval changes of the same nodule in sequential 3D ROIs and capture spatial dependencies of nodule regions and the current 3D ROI. According to the clinical diagnosis routine, we employ hierarchical loss to pay more attention to growing nodules. The extensive experiments on our organized dataset demonstrate the advantage of our proposed method. We also conduct experiments on an in-house dataset to evaluate the clinical utility of our method by comparing it against skilled clinicians. STM code and NLSTt dataset are available at https://github.com/liaw05/STMixer. |
Keywords | |
SUSTech Authorship | Corresponding
|
Language | English
|
URL | [Source Record] |
Indexed By | |
Funding Project | Guangdong Provincial Department of Education[2020ZDZX3043]
; Shenzhen Natural Science Fund["JCYJ20200109140820699","20200925174052004"]
|
WOS Research Area | Computer Science
; Neurosciences & Neurology
; Radiology, Nuclear Medicine & Medical Imaging
|
WOS Subject | Computer Science, Interdisciplinary Applications
; Neuroimaging
; Radiology, Nuclear Medicine & Medical Imaging
|
WOS Accession No | WOS:000867524300046
|
Scopus EID | 2-s2.0-85138797394
|
Data Source | Scopus
|
Citation statistics |
Cited Times [WOS]:1
|
Document Type | Conference paper |
Identifier | http://kc.sustech.edu.cn/handle/2SGJ60CL/402743 |
Department | Research Institute of Trustworthy Autonomous Systems |
Affiliation | 1.School of Computer Science and Technology,Harbin Institute of Technology,Harbin,China 2.CVTE Research,Guangzhou,China 3.Research Institute of Trustworthy Autonomous Systems,Southern University of Science and Technology,Shenzhen,China 4.School of Computer,Guangdong University of Technology,Guangzhou,China 5.Yibicom Health Management Center,CVTE,Guangzhou,China |
First Author Affilication | Research Institute of Trustworthy Autonomous Systems |
Corresponding Author Affilication | Research Institute of Trustworthy Autonomous Systems |
Recommended Citation GB/T 7714 |
Fang,Jiansheng,Wang,Jingwen,Li,Anwei,et al. Siamese Encoder-based Spatial-Temporal Mixer for Growth Trend Prediction of Lung Nodules on CT Scans[C]. GEWERBESTRASSE 11, CHAM, CH-6330, SWITZERLAND:SPRINGER INTERNATIONAL PUBLISHING AG,2022:484-494.
|
Files in This Item: | There are no files associated with this item. |
|
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment