中文版 | English
Title

Domain Adaptive Retinal Vessel Segmentation Guided by High-frequency Component

Author
Corresponding AuthorLi,Heng
DOI
Publication Years
2022
Conference Name
9th International Workshop on Ophthalmic Medical Image Analysis (OMIA)
ISSN
0302-9743
EISSN
1611-3349
ISBN
978-3-031-16524-5
Source Title
Volume
13576 LNCS
Pages
115-124
Conference Date
SEP 22, 2022
Conference Place
null,Singapore,SINGAPORE
Publication Place
GEWERBESTRASSE 11, CHAM, CH-6330, SWITZERLAND
Publisher
Abstract
The morphological structure of retinal fundus blood vessels is of great significance for medical diagnosis, thus the automatic retinal vessel segmentation algorithm has become one of the research hotspots in the field of medical image processing. However, there are still several unsolved difficulties in this task: the existed methods are too sensitive to the low-frequency noise in the fundus images, and there are few annotated data sets available, and meanwhile, the retinal images of different datasets vary greatly. To solve the above problems, we propose a domain adaptive vessel segmentation algorithm with multiple image entrances called MIUnet, which is robust to the etiological noises and domain shift between diverse datasets. We apply Fourier domain adaptation and the high-frequency component filtering modules to transform the raw images into two styles, and simultaneously reduce the discrepancy between the source domain and target domain retinal images. After that, images produced by the two modules are fed into a multi-input deep segmentation model, and the full utilization of features from different modalities is ensured by the deep supervision mechanism. Experiments prove that, compared with other segmentation methods, the MIUnet has better performances in cross-domain experiments, where the IoU reaches 63% when trained on ARIA dataset and tested on the DRIVE dataset and 53% in the opposite direction.
Keywords
SUSTech Authorship
First ; Corresponding
Language
English
URL[Source Record]
Indexed By
WOS Research Area
Ophthalmology ; Radiology, Nuclear Medicine & Medical Imaging
WOS Subject
Ophthalmology ; Radiology, Nuclear Medicine & Medical Imaging
WOS Accession No
WOS:000869749600012
Scopus EID
2-s2.0-85138772563
Data Source
Scopus
Citation statistics
Cited Times [WOS]:0
Document TypeConference paper
Identifierhttp://kc.sustech.edu.cn/handle/2SGJ60CL/402750
DepartmentDepartment of Computer Science and Engineering
Affiliation
1.Guangdong Provincial Key Laboratory of Brain-inspired Intelligent Computation,Southern University of Science and Technology,Shenzhen,China
2.Department of Computer Science and Engineering,Southern University of Science and Technology,Shenzhen,China
3.The School of Computer and Communication Engineering,University of Science and Technology,Beijing,China
4.Singapore Eye Research Institute,Singapore National Eye Centre,Singapore,Singapore
First Author AffilicationSouthern University of Science and Technology;  Department of Computer Science and Engineering
Corresponding Author AffilicationDepartment of Computer Science and Engineering
First Author's First AffilicationSouthern University of Science and Technology
Recommended Citation
GB/T 7714
Li,Haojin,Li,Heng,Qiu,Zhongxi,et al. Domain Adaptive Retinal Vessel Segmentation Guided by High-frequency Component[C]. GEWERBESTRASSE 11, CHAM, CH-6330, SWITZERLAND:SPRINGER INTERNATIONAL PUBLISHING AG,2022:115-124.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Export to Excel
Export to Csv
Altmetrics Score
Google Scholar
Similar articles in Google Scholar
[Li,Haojin]'s Articles
[Li,Heng]'s Articles
[Qiu,Zhongxi]'s Articles
Baidu Scholar
Similar articles in Baidu Scholar
[Li,Haojin]'s Articles
[Li,Heng]'s Articles
[Qiu,Zhongxi]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Li,Haojin]'s Articles
[Li,Heng]'s Articles
[Qiu,Zhongxi]'s Articles
Terms of Use
No data!
Social Bookmark/Share
No comment.

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.