Reference Class-Based Improvement of Object Detection Accuracy

Raegeun Park, Jaechoon Jo

Abstract


To date, the Frames Per Second (FPS) and accuracy of object detection based on deep learning have made rapid progress. However, the accuracy is limited by issues such as false positive (FP) cases. FP cases can trigger malfunctions in applications requiring high accuracy, such as in autonomous vehicles, where it is essential to ensure driver safety when malfunctions occur. To reduce the occurrences of FP cases, we conducted an experiment to derive the association by separately detecting a highly relevant element called a reference class, in addition to the target class to be detected. To measure the association, we obtained the integrated association by first finding the associations between the bounding boxes of the target and reference classes. Then we generated a reference class-based model by applying the integrated association to a trained model. The reference class-based model achieved approximately 15% higher accuracy than the trained model at iteration 1,000. Besides, the proposed model reduced the FP cases to approximately half of the 18.964% in the conventional method; the FP reduction through an increase in iteration was only 11.008%. The reference class can be applied in various fields, such as security and autonomous vehicle technology. It can be used to reduce the FP cases and improve the accuracy performance limits in object detection. Furthermore, it is possible to reduce the cost of reinforcing the training dataset and using high-performance hardware, and the time cost of increasing training numbers.


Keywords


reference class; target class; FP case; association; improvement of accuracy performance.

Full Text:

PDF

References


A. Krizhevsky, I. Sutskever and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,†In Advances in neural information processing systems, pp. 1097-1105, 2012

E. Karami, S. Prasad and M. Shehata, “Image matching using SIFT, SURF, BRIEF and ORB: performance comparison for distorted images,†arXiv preprint arXiv:1710.02726, Oct. 2017

J. Redmon and A. Farhadi, “YOLO9000: Better, Faster, Stronger,†The IEEE conference on computer vision and pattern recognition, pp. 7263-7271, 2017

J. Redmon and A. Farhadi, “YOLOv3: An Incremental Improvement,†arXiv:1804.02767, 2018

YOLO. (2020), YOLO:Real-Time Object Detection. [Online]. Available: https://pjreddie.com/darknet/yolo/

T. Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár and C. L. Zitnick, “Microsoft COCO: Common Objects in Contextâ€, European conference on computer vision. Springer, pp. 740-755, 2014

Y. Long, Y. Gong, Z. Xiao and Q. Liu, “Accurate object localization in remote sensing images based on convolutional neural networks,†IEEE Transactions on Geoscience and Remote Sensing, 55(5), pp. 2486-2498, Jan. 2017

X. Wang, A. Shrivastava and A. Gupta, “A-fast-rcnn: Hard positive generation via adversary for object detection,†the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2606-2615, 2017

B. Singh and L. S. Davis, “An analysis of scale invariance in object detection snip,†the IEEE conference on computer vision and pattern recognition, pp. 3578-3587, 2018

Z. Cai and N. Vasconcelos, “Cascade r-cnn: Delving into high quality object detection,†the IEEE conference on computer vision and pattern recognition, pp. 6154-6162, 2018

Y. Chen, W. Li, C. Sakaridis, D. Dai and L. V. Gool, “Domain adaptive faster r-cnn for object detection in the wild,†the IEEE conference on computer vision and pattern recognition, pp. 3339-3348, 2018

J. Jeong, H. Park and N. Kwak, “Enhancement of SSD by concatenating feature maps for object detection,†arXiv preprint arXiv:1705.09587, May. 2017

X. Sun, P. Wu and S. C. H. Hoi, “Face detection using deep learning: An improved faster RCNN approach,†Neurocomputing, 299, 42-50, 2018

X. Zhu, Y. Wang, J. Dai, L. Yuan and Y. Wei, “Flow-guided feature aggregation for video object detection,†the IEEE International Conference on Computer Vision, pp. 408-417, 2017

T. Y. Lin, P. Goyal, R. Girshick, K. He and P. Dollár, “Focal loss for dense object detection,†the IEEE international conference on computer vision, pp. 2980-2988, 2017

T. Y. Lin and S. Maji, “Improved bilinear pooling with cnns,†arXiv preprint arXiv:1707.06772, 2017

L. Tychsen-Smith and L. Petersson, “Improving object localization with fitness nms and bounded iou loss,†the IEEE conference on computer vision and pattern recognition, pp. 6877-6885, 2018

X. Wang, X. Hua, F. Xiao, Y. Li, X. Hu and P. Sun, “Multi-object detection in traffic scenes based on improved SSD,†Electronics, 7(11), 302, 2018

G. Bertasius, L. Torresani and J. Shi, “Object detection in video with spatiotemporal sampling networks,†the European Conference on Computer Vision (ECCV), pp. 331-346, 2018

J. Li, X. Liang, Y. Wei, T. Xu, J. Feng and S. Yan, “Perceptual generative adversarial networks for small object detection,†the IEEE conference on computer vision and pattern recognition, pp. 1222-1230, 2017

T. Kong, F. Sun, A. Yao, H. Liu, M. Lu and Y. Chen, “Ron: Reverse connection with objectness prior networks for object detection,†the IEEE conference on computer vision and pattern recognition, pp. 5936-5944, 2017

Y. Li, Y. Chen, N. Wang and Z. Zhang, “Scale-aware trident networks for object detection,†the IEEE International Conference on Computer Vision, pp. 6054-6063, 2019

P. Zhou, B. Ni, C. Geng, J. Hu and Y. Xu, “Scale-transferrable object detection,†the IEEE conference on computer vision and pattern recognition, pp. 528-537, 2018

Z. Zhang, S. Qiao, C. Xie, W. Shen, B. Wang and A. L. Yuille, “Single-shot object detection with enriched semantics,†the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5813-5821, 2018

S. Zhang, L. Wen, X. Bian, Z. Lei and S. Z. Li, “Single-shot refinement neural network for object detection,†the IEEE conference on computer vision and pattern recognition, pp. 4203-4212, 2018

N. Bodla, B. Singh, R. Chellappa and L. S. Davis, “Soft-NMS--improving object detection with one line of code,†the IEEE international conference on computer vision, pp. 5561-5569, 2017

K. Kang, H. Li, J. Yan, X. Zeng, B. Yang, T. Xiao, C. Zhang, Z. Wang, R. Wang, X. Wang and W. Ouyang, “T-cnn: Tubelets with convolutional neural networks for object detection from videos,†IEEE Transactions on Circuits and Systems for Video Technology, 28(10), pp. 2896-2907, 2017

R. G. Park, H. J. Yun, E. G. Han, S. W. Kang, J. H. Park, E. J. Lee, D. H. Jeon, K. Y. Jung, S. B. Cho and T. K. Kang, “A Study on Countermeasures for ADAS Malfunction Based on YOLOâ€, Korean Institute of Electrical Engineers(KIEE) Conference, pp.188-190, Nov. 2019

T. Karras, S. Laine and T. Aila, “A style-based generator architecture for generative adversarial networks,†The IEEE Conference on Computer Vision and Pattern Recognition, pp. 4401-4410, 2019

N. Dalal and B. Triggs, “Histograms of oriented gradients for human detection,†The IEEE computer society conference on computer vision and pattern recognition, pp. 886-893, Jun. 2005

M. Everingham, A. Zisserman, C. K. I. Williams, L. V. Gool, M. Allan, C. M. Bishop, O. Chapelle, N. Dalal, T. Deselaers, G. Dork´o, S. Duffner, J. Eichhorn, J. D. R. Farquhar, M. Fritz, C. Garcia, T. Griffiths, F. Jurie, D. Keysers, M. Koskela, J. Laaksonen, D. Larlus and B. Leibe, “The 2005 PASCAL Visual Object Classes Challengeâ€, In Machine Learning Challenges Workshop, pp. 117-176, Apr. 2005

V. S. Rotenberg, “Moravec’s paradox: consideration in the context of two brain hemisphere functions,†Activitas Nervosa Superior, 55(3), pp. 108-111, 2013

(2020) Joint Base Langley-Eustis website. [Online]. Available: https://www.jble.af.mil/News/Photos/igphoto/2000197216/




DOI: http://dx.doi.org/10.18517/ijaseit.10.4.12792

Refbacks

  • There are currently no refbacks.



Published by INSIGHT - Indonesian Society for Knowledge and Human Development