Skip to main content
Log in

Lidar Point Semantic Segmentation Using Dual Attention Mechanism

  • Published:
Journal of Russian Laser Research Aims and scope

Abstract

A large amount of environmental point cloud data can be provided by light detection and ranging (LiDAR) sensors. The raw points collected by the LiDAR is disordered and unstructured. It is a challenge to design algorithms to extract features from the raw points. In this paper, we propose the LiDAR point semantic segmentation net (LPSS Net), which is a dual-attention mechanism point cloud segmentation algorithm. First, the LPSS Net extracts point cloud features from the raw points, which uses the self-attention mechanism in the transformer mechanism. Second, in order to suppress irrelevant information in the features and focus on essential information, we propose a novel strategy, which designs a 3D channel attention mechanism in the encoding part. Finally, we demonstrate the applicability of the algorithm by extensive experiments conducted on Semantic3D and SemanticKITTI. A high OA of 90.6% and a mIoU of 71.5% on the Semantic3D data set indicate the feasibility of the algorithm.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. C. R. Qi, H. Su, K. Su, et al., “Pointnet: Deep learning on point sets for 3D classification and segmentation,” in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017), pp. 652–660.

  2. C. R. Qi, L. Yi, H. Su, et al., “Pointnet++: Deep hierarchical feature learning on point sets in a metric space,” in: Advances in Neural Information Processing Systems, 30, Curran Associates (2017); ISBN: 9781510860964

  3. H. Zhao, L. Jiang, J. Jia, et al., “Point transformer,” in: Proceedings of the IEEE/CVF International Conference on Computer Vision (2021), pp. 16259–16268.

  4. T. Hackel, N. Savinov, L. Ladicky, et al., Semantic3D.net: A new large-scale point cloud classification benchmark, arXiv: 1704.03847 (2017).

  5. J. Behley, M. Garbade, A. Milioto, et al., “SemanticKITTI: A dataset for semantic scene understanding of LiDAR sequences,” in: Proceedings of the IEEE/CVF International Conference on Computer Vision (2019), pp. 9297–9307.

  6. L. Ma, J. Stckler, C. Kerl, et al., “Multi-view deep learning for consistent semantic mapping with RGB-D cameras,” in: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE (2017), pp. 598–605.

  7. D. Maturana and S. Scherer, “Voxnet: A 3D convolutional neural network for real-time object recognition,” in: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE (2015), pp. 922–928.

  8. G. Qian, Y. Li, H. Peng, et al., PointNeXt: Revisiting PointNet++ with improved training and scaling strategies, arXiv: 2206.04670 (2022).

  9. A. Vaswani, N. Shazeer, N. Parmar, et al., “Attention is all you need,” in: Advances in Neural Information Processing Systems, 30, Curran Associates (2017); ISBN: 9781510860964

  10. J. Fu, J. Liu, J. Jiang, et al., IEEE Trans. Neural Netw. Learn. Syst., 32, 2547 (2020).

    Article  Google Scholar 

  11. L. C. Chen, Y. Zhu, G. Papandreou, et al., “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in: Proceedings of the European Conference on Computer Vision (ECCV) (2018), pp. 801–818.

  12. K. Han, A. Xiao, E. Wu, et al., “Transformer in transformer,” in: Advances in Neural Information Processing Systems, 34, Curran Associates (2021); ISBN: 9781713845393

  13. W. Wu, Z. Qi, and L. Fuxin, “PointConv: Deep convolutional networks on 3D point clouds, ” in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2019), pp. 9621–9630.

  14. Y. Zhou, Q. Sun, J. Meng, et al., J. Opt. Soc. Am. A: Opt. Image Sci. Vis., 38, 1194 (2021).

  15. Q. Hu, B. Yang, L. Xie, et al., “Randla-net: Efficient semantic segmentation of large-scale point clouds,” in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2020), pp. 11108–11117.

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shifeng Wang.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, H., Zhou, Y., Chen, T. et al. Lidar Point Semantic Segmentation Using Dual Attention Mechanism. J Russ Laser Res 44, 224–234 (2023). https://doi.org/10.1007/s10946-023-10127-9

Download citation

  • Received:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10946-023-10127-9

Keywords

Navigation