Ajou University repository

3D Directional Encoding for Point Cloud Analysisoa mark
Citations

SCOPUS

0

Citation Export

Publication Year
2024-01-01
Journal
IEEE Access
Publisher
Institute of Electrical and Electronics Engineers Inc.
Citation
IEEE Access, Vol.12, pp.144533-144543
Keyword
Classificationdeep learningdirectional feature extractionefficient neural networkpoint cloudsegmentation
Mesh Keyword
Deep learningDirectional features extractionsEfficient neural networkEncodingsLocal featureNeural-networksPoint featuresPoint-cloudsSegmentationSimple++
All Science Classification Codes (ASJC)
Computer Science (all)Materials Science (all)Engineering (all)
Abstract
Extracting informative local features in point clouds is crucial for accurately understanding spatial information inside 3D point data. Previous works utilize either complex network designs or simple multi-layer perceptrons (MLP) to extract the local features. However, complex networks often incur high computational cost, whereas simple MLP may struggle to capture the spatial relations among local points effectively. These challenges limit their scalability to delicate and real-time tasks, such as autonomous driving and robot navigation. To address these challenges, we propose a novel 3D Directional Encoding Network (3D-DENet) capable of effectively encoding spatial relations with low computational cost. 3D-DENet extracts spatial and point features separately. The key component of 3D-DENet for spatial feature extraction is Directional Encoding (DE), which encodes the cosine similarity between direction vectors of local points and trainable direction vectors. To extract point features, we also propose Local Point Feature Multi-Aggregation (LPFMA), which integrates various aspects of local point features using diverse aggregation functions. By leveraging DE and LPFMA in a hierarchical structure, 3D-DENet efficiently captures both detailed spatial and high-level semantic features from point clouds. Experiments show that 3D-DENet is effective and efficient in classification and segmentation tasks. In particular, 3D-DENet achieves an overall accuracy of 90.7% and a mean accuracy of 90.1% on ScanObjectNN, outperforming the current state-of-the-art method while using only 47% floating point operations.
ISSN
2169-3536
Language
eng
URI
https://aurora.ajou.ac.kr/handle/2018.oak/38084
https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85205858678&origin=inward
DOI
https://doi.org/10.1109/access.2024.3472301
Journal URL
http://ieeexplore.ieee.org/xpl/RecentIssue.jsp?punumber=6287639
Type
Article
Funding
This work was supported in part by the BK21 FOUR Program of the Education and Research Program for Future ICT Pioneers, Seoul National University in 2024, and Institute of Information and Communications Technology Planning and Evaluation (IITP) grant funded by the Korea government (MSIT) under Grant RS-2022-00207391 and in part by the Development of Hashgraph-based Blockchain Enhancement Scheme and Implementation of Testbed for Autonomous Driving.
Show full item record

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Lee, Sang Hyun Image
Lee, Sang Hyun이상현
Department of Mobility Engineering
Read More

Total Views & Downloads

File Download

  • There are no files associated with this item.