Multimodal Fusion and Data Augmentation for 3D Semantic Segmentation

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 66
  • Download : 0
Since modern autonomous driving (AD) platforms offer a variety of sensors, it is intuitive to leverage complementary data from multimodal sensors to produce reliable 3D semantic segmentation. However, due to the information loss and the sub-optimized fusion in multimodal fusion methods, LiDAR-only methods currently occupy the top positions in the leaderboard of datasets. In this paper, we focus on two aspects to improve the LiDAR-camera fusion semantic segmentation performance, namely data augmentation and fusion strategy. First, we propose an novel data augmentation by refining point-image patches. Second, we design an attention fusion block for the dual-branch segmentation network by considering the modality gap between LiDAR and RGB camera. Experiments on nuScences indicate that our proposed method outperforms the baseline methods on key classes.
Publisher
IEEE Computer Society
Issue Date
2022-11
Language
English
Citation

22nd International Conference on Control, Automation and Systems, ICCAS 2022, pp.1143 - 1148

ISSN
1598-7833
DOI
10.23919/ICCAS55662.2022.10003729
URI
http://hdl.handle.net/10203/305120
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0