Abstract

Deep learning models used to process 3D point cloud data exhibit a vulnerability to adversarial examples, which can result in potential misclassification. The research community has pivoted towards applying traditional optimization methods to generate adversarial examples specifically for the point cloud domain. However, these traditional optimization-based adversarial approaches come with their own set of limitations, including inconsistent point importance, non-optimal multi-label optimization, inflexible adversarial and perceptibility losses, and difficulty in constructing perceivable adversarial examples in physical environments. To surmount these challenges, we have introduced two novel techniques: a Local Aggressive Adversarial Attack (L3A) for simulated environments and a Normal Vector Metric Attack (NVMA) for physical environments. The L3A operates by enhancing the cost-effectiveness ratio between the attack and perceptibility aspects. This enhancement is realized through a combination of strategies, including local salient points, modeling example preference, and adjusting the goals for attack potency and indistinguishability. The NVMA generates physically printable adversarial point clouds by maintaining consistency in high-frequency regions between clean and adversarial examples. Our methods exhibit superior performance when tested against state-of-the-art point cloud deep learning models on benchmark datasets and physical models. This contribution enhances our understanding of adversarial attacks and defenses within the realm of 3D point cloud data. Our code is available at https://github.com/Chenfeng1271/L3A.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call