Object detection in seriously degraded images with unbalanced training samples
CSTR:
Author:
Affiliation:

Collage of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China

  • Article
  • | |
  • Metrics
  • |
  • Reference [20]
  • |
  • Related [20]
  • | | |
  • Comments
    Abstract:

    Uncertain environments, especially uneven lighting and shadows, can degrade an image, which causes a great negative impact on object detection. Moreover, unbalanced training samples can cause overfitting problem. Since available data that is collected at night is much rarer than that collected in the daytime, the nighttime detection effect will be relatively poor. In this paper, we propose a novel data augmentation method named Mask Augmentation, which reduces the brightness and contrast of objects, and also weakens the edge of objects to simulate the degraded scene. In addition, we propose a new architecture, by adding a classification loss branch and a feature extraction module named Multi-Feature Attention Module, which combines the attention mechanism and feature fusion on the basis of Darknet-53. This architecture makes the features extracted in daytime and nighttime images distinguishable. We also increase the loss weight of nighttime images during the training process. We achieved 78.68% mAP on nighttime detection and 73.14% mAP on daytime detection. Compared with other models, our method greatly improves the accuracy of nighttime detection, and also performs satisfactorily on daytime detection. We deployed our model on an intelligent garbage collection robot for real-time detection, which implements automatic picking at night and assists cleaning staff during the day.

    Reference
    [1] Ren S, He K, Girshick R and J. Sun, IEEE Transactions on Pattern Analysis and Machine Intelligence 39, 1137 (2015).
    [2] Cai Z and Vasconcelos N, Cascade R-CNN:Delving into High Quality Object Detection, arXiv:1712.00726v1, 2018.
    [3] Redmon J and Farhadi A, YOLOv3:An Incremental Improvement, arXiv:1804.02767, 2018.
    [4] Tian Z, Shen C, Chen H and He T, Fcos:Fully Convolutional One-Stage Object Detection, IEEE International Conference on Computer Vision, 9627 (2019).
    [5] Kong T, Sun F, Liu H, Jiang Y and Shi J, Foveabox:Beyond Anchor-Based Object Detector, arXiv:1904.03797, 2019.
    [6] Yang Z, Liu S, Hu H, Wang L and Lin S, Reppoints:Point Set Representation for Object Detection, IEEE International Conference on Computer Vision, 9657 (2019).
    [7] Proen?a P F and Sim?es P, TACO:Trash Annotations in Context for Litter Detection, arXiv:2003.06975, 2020.
    [8] Lowe D G, International Journal of Computer Vision 60, 91 (2004).
    [9] Platt J, Sequential Minimal Optimization:A Fast Algorithm for Training Support Vector Machines, 1998.
    [10] Girshick R, Donahue J, Darrell T and Malik J, Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation, IEEE Conference on computer vision and pattern recognition, 580 (2014).
    [11] Girshick R, Fast R-CNN, IEEE International Conference on Computer Vision, 1440 (2015).
    [12] Redmon J, Divvala S, Girshick R and Farhadi A, You Only Look Once:Uni?ed, Real-Time Object Detection, IEEE Conference on Computer Vision and Pattern Recognition, 779 (2016).
    [13] Redmon J and Farhadi A, YOLO9000:Better, Faster, Stronger, IEEE Conference on Computer Vision and Pattern Recognition, 7263 (2017).
    [14] Lin T Y, Goyal P, Girshick R, He K and Dollár P, Focal Loss for Dense Object Detection, IEEE International Conference on Computer Vision, 2980 (2017).
    [15] Zhou X, Wang D and Kr?henbühl P, Objects as Points, Computer Vision and Pattern Recognition, arXiv:1904.07850, 2019.
    [16] Fan D P, Ji G P, Sun G, Cheng MM and L Shao, Camouflaged Object Detection, IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
    [17] Fan Q, Zhuo W, Tang C K and Tai Y W, Few-Shot Object Detection with Attention-RPN and Multi-Relation Detector, IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
    [18] Cao J, Cholakkal H, Anwer R M, Khan F S and Shao L, D2Det:Towards High Quality Object Detection and Instance Segmentation, IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
    [19] Ioffe S and Szegedy C, Batch Normalization:Accelerating Deep Network Training by Reducing Internal Covariate Shift, arXiv:1502.03167, 2015.
    [20] Maas A L, Hannun A Y and Ng A Y, Recti?er Nonlinearities Improve Neural Network Acoustic Models, ICML Workshop on Deep Learning for Audio, Speech and Language Processing, 3 (2013).
    Cited by
    Comments
    Comments
    分享到微博
    Submit
Get Citation

LIU Sheng, SHEN Jiayu, HUANG Shengyue. Object detection in seriously degraded images with unbalanced training samples[J]. Optoelectronics Letters,2021,17(9):564-571

Copy
Share
Article Metrics
  • Abstract:861
  • PDF: 86
  • HTML: 0
  • Cited by: 0
History
  • Received:January 12,2021
  • Revised:February 25,2021
  • Online: July 09,2021
Article QR Code