Dual-Perspective Knowledge Enrichment for Semi-supervised 3D Object Detection

Page view(s)
8
Checked on Mar 01, 2025
Dual-Perspective Knowledge Enrichment for Semi-supervised 3D Object Detection
Title:
Dual-Perspective Knowledge Enrichment for Semi-supervised 3D Object Detection
Journal Title:
Proceedings of the AAAI Conference on Artificial Intelligence 2024
Publication Date:
25 March 2024
Citation:
Han, Y., Zhao, N., Chen, W., Ma, K. T., & Zhang, H. (2024). Dual-Perspective Knowledge Enrichment for Semi-supervised 3D Object Detection. Proceedings of the AAAI Conference on Artificial Intelligence, 38(3), 2049–2057. https://doi.org/10.1609/aaai.v38i3.27976
Abstract:
Semi-supervised 3D object detection is a promising yet under-explored direction to reduce data annotation costs, especially for cluttered indoor scenes. A few prior works, such as SESS and 3DIoUMatch, attempt to solve this task by utilizing a teacher model to generate pseudo-labels for unlabeled samples. However, the availability of unlabeled samples in the 3D domain is relatively limited compared to its 2D counterpart due to the greater effort required to collect 3D data. Moreover, the loose consistency regularization in SESS and restricted pseudo-label selection strategy in 3DIoUMatch lead to either low-quality supervision or a limited amount of pseudo labels. To address these issues, we present a novel Dual-Perspective Knowledge Enrichment approach named DPKE for semi-supervised 3D object detection. Our DPKE enriches the knowledge of limited training data, particularly unlabeled data, from two perspectives: data-perspective and feature-perspective. Specifically, from the data-perspective, we propose a class-probabilistic data augmentation method that augments the input data with additional instances based on the varying distribution of class probabilities. Our DPKE achieves feature-perspective knowledge enrichment by designing a geometry-aware feature matching method that regularizes feature-level similarity between object proposals from the student and teacher models. Extensive experiments on the two benchmark datasets demonstrate that our DPKE achieves superior performance over existing state-of-the-art approaches under various label ratio conditions. The source code and models will be made available to the public.
License type:
Publisher Copyright
Funding Info:
This research / project is supported by the A*STAR - MTC Programmatic Fund
Grant Reference no. : M23L7b0021

This research / project is supported by the National Research Foundation - AI Singapore Programme
Grant Reference no. : AISG2-RP-2021-022

This research / project is supported by the Hyundai - Hyundai Research Grant
Grant Reference no. : 04OIS000257C130
Description:
This material may not be retransmitted or redistributed without permission in writing from The Association for the Advancement of Artificial Intelligence. Permission to use document is granted, provided that (1) the copyright notice appears in all copies and that both the copyright notice and this permission notice appear, (2) use of such documents is for personal use only, and will not be copied or posted on any network computer or broadcast in any media, and (3) no modifications of any documents are made.
ISSN:
2374-3468
2159-5399