2023 Vol.6 Aug.N04 |
---|
|
Reference: [1] GIRSHICK ROSS. Fast R-CNN[C]. In Proceedings of the IEEE International Conference on Computer Vision, 2015: 1440-1448. [2] BOCHKOVSKIY A, WANG C Y, LIAO H Y M. Yolov4: Optimal speed and accuracy of object detection[J]. arXiv preprint, 2020. [3] GE Z, LIU S, WANG F, et al. YoloX: Exceeding yolo series in 2021[J]. arXiv preprint, 2021. [4] DU D, ZHU P, WEN L, et al. VisDrone-SOT2019: The Vision meets Drone Single Object Tracking Challenge Results[C]. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019: 199-212. [5] CAI Z, VASCONCELOS N. Cascade R-CNN: Delving into high quality object detection[C]. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2018: 6154-6162. [6] LIN T Y, GOYAL P, GIRSHICK R, et al. Focal loss for dense object detection[C]. In Proceedings of the IEEE International Conference on Computer Vision, 2017: 2980-2988. [7] LIN T Y, DOLLÁR P, GIRSHICK R, et al. Feature Pyramid Networks for Object Detection[C]. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017: 2117-2125. [8] LIU S, QI L, QIN H, et al. Path aggregation network for instance segmentation[C]. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018: 8759-8768. |
Tsuruta Institute of Medical Information Technology
Address:[502,5-47-6], Tsuyama, Tsukuba, Saitama, Japan TEL:008148-28809 fax:008148-28808 Japan,Email:jpciams@hotmail.com,2019-09-16