Title | Discrete time convolution for fast event-based stereo |
Author | |
Corresponding Author | Luziwei Leng |
Joint first author | Kaixuan Zhang; Luziwei Leng |
DOI | |
Publication Years | 2022
|
Conference Name | Conference on Computer Vision and Pattern Recognition (CVPR)
|
ISSN | 1063-6919
|
ISBN | 978-1-6654-6947-0
|
Source Title | |
Volume | 2022-June
|
Pages | 8666-8676
|
Conference Date | 18-24 June 2022
|
Conference Place | New Orleans, LA, USA
|
Publication Place | 10662 LOS VAQUEROS CIRCLE, PO BOX 3014, LOS ALAMITOS, CA 90720-1264 USA
|
Publisher | |
Abstract | Inspired by biological retina, dynamical vision sensor transmits events of instantaneous changes of pixel intensity, giving it a series of advantages over traditional frame-based camera, such as high dynamical range, high temporal resolution and low power consumption. However, extracting information from highly asynchronous event data is a challenging task. Inspired by continuous dynamics of biological neuron models, we propose a novel encoding method for sparse events - continuous time convolution (CTC) - which learns to model the spatial feature of the data with intrinsic dynamics. Adopting channel-wise parameterization, temporal dynamics of the model is synchronized on the same feature map and diverges across different ones, enabling it to embed data in a variety of temporal scales. Abstracted from CTC, we further develop discrete time convolution (DTC) which accelerates the process with lower computational cost. We apply these methods to event-based multi-view stereo matching where they surpass state-of-the-art methods on benchmark criteria of the MVSEC dataset. Spatially sparse event data often leads to inaccurate estimation of edges and local contours. To address this problem, we propose a dual-path architecture in which the feature map is complemented by underlying edge information from original events extracted with spatially-adaptive denormalization. We demonstrate the superiority of our model in terms of speed (up to 110 FPS), accuracy and robustness, showing a great potential for real-time fast depth estimation. Finally, we perform experiments on the recent DSEC dataset to demonstrate the general usage of our model. |
Keywords | |
SUSTech Authorship | First
|
Language | English
|
URL | [Source Record] |
Indexed By | |
Funding Project | National Key Research and Development Program of China[2021YFF1200800]
|
WOS Research Area | Computer Science
; Imaging Science & Photographic Technology
|
WOS Subject | Computer Science, Artificial Intelligence
; Imaging Science & Photographic Technology
|
WOS Accession No | WOS:000870759101070
|
EI Accession Number | 20224613120585
|
Data Source | IEEE
|
PDF url | https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9878658 |
Citation statistics |
Cited Times [WOS]:0
|
Document Type | Conference paper |
Identifier | http://kc.sustech.edu.cn/handle/2SGJ60CL/406479 |
Department | Department of Computer Science and Engineering 工学院_电子与电气工程系 |
Affiliation | 1.Department of Computer Science and Engineering, Southern University of Science and Technology, China 2.Department of Electrical and Electronic Engineering, Southern University of Science and Technology, China 3.ACS Lab, Huawei Technologies, Shenzhen, China 4.Peng Cheng Lab, Shenzhen, China |
First Author Affilication | Department of Computer Science and Engineering |
First Author's First Affilication | Department of Computer Science and Engineering |
Recommended Citation GB/T 7714 |
Kaixuan Zhang,Kaiwei Che,Jianguo Zhang,et al. Discrete time convolution for fast event-based stereo[C]. 10662 LOS VAQUEROS CIRCLE, PO BOX 3014, LOS ALAMITOS, CA 90720-1264 USA:IEEE COMPUTER SOC,2022:8666-8676.
|
Files in This Item: | ||||||
File Name/Size | DocType | Version | Access | License | ||
Zhang_Discrete_Time_(2966KB) | Restricted Access | -- |
|
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment