The rapid growth of real-time huge data capturing has pushed the deep
learning and data analytic computing to the edge systems. Real-time object
recognition on the edge is one of the representative deep neural network (DNN)
powered edge systems for real-world mission-critical applications, such as
autonomous driving and augmented reality. While DNN powered object detection
edge systems celebrate many life-enriching opportunities, they also open doors
for misuse and abuse. This paper presents three Targeted adversarial Objectness
Gradient attacks, coined as TOG, which can cause the state-of-the-art deep
object detection networks to suffer from object-vanishing, object-fabrication,
and object-mislabeling attacks. We also present a universal objectness gradient
attack to use adversarial transferability for black-box attacks, which is
effective on any inputs with negligible attack time cost, low human
perceptibility, and particularly detrimental to object detection edge systems.
We report our experimental measurements using two benchmark datasets (PASCAL
VOC and MS COCO) on two state-of-the-art detection algorithms (YOLO and SSD).
The results demonstrate serious adversarial vulnerabilities and the compelling
need for developing robust object detection systems.