|
李 佳, 刘 芳, 李迪强, 等. 2017. 基于红外相机监测分析的红腹角雉日活动节律. 林业科学, 53(7): 170−176.
|
|
Li J, Liu F, Li D Q, et al. 2017. Daily activity rhythm of Temminick’s Tragopan (Trgopan temminckii) based on infrared camera monitoring. Journal of Computer Applications. [in Chinese]
|
|
李 果, 李俊生, 关 潇, 等. 2014. 生物多样性监测技术手册. 北京: 中国环境科学出版社.
|
|
Li G, Li J S, Guan X, et al. 2014. Biodiversity monitoring technical manuals. Beijing: China Environmental Science Press. [in Chinese]
|
|
Ali H, Vishwesh N, Yucheng T, et al. 2022. Swin unetr: swin transformers for semantic segmentation of brain tumors in mri images. International MICCAI Brainlesion Workshop, Springer, 272–284.
|
|
Bochkovskiy A, Wang C Y, Liao H Y M. 2020. YOLOv4: optimal speed and accuracy of object detection. arXiv Preprint arXiv: 2004.10934.
|
|
Chen R, Little R, Mihaylova L, et al. Forest wildlife surveillance using deep learning methods. Ecology and Evolution, 2019, 9 (17): 9453- 9466.
doi: 10.1002/ece3.5410
|
|
Chen G, Han T X, He Z, et al. 2014. Deep convolutional neural network based species recognition for wild animal monitoring. 2014 IEEE International Conference on Image Processing (ICIP), IEEE, 858−862.
|
|
Carion N, Massa F, Synnaeve G, et al. 2020. End-to-end object detection with transformers. European Conference on Computer Vision (ECCV), 12346: 213–229.
|
|
DeVries T, Taylor G W. 2017. Improved regularization of convolutional neural networks with cutout. arXiv Preprint arXiv: 1708.04552.
|
|
Dosovitskiy A, Beyer L, Kolesnikov A, et al. 2020. An image is worth 16×16 words: transformers for image recognition at scale. arXiv Preprint arXiv: 2010.11929.
|
|
Girshick R. 2015. Fast R-CNN. Proceedings of the IEEE International Conference on Computer Vision, 1440−1448.
|
|
Girshick R, Donahue J, Darrell T, et al. 2014. Rich feature hierarchies for accurate object detection and semantic segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 580−587.
|
|
He K, Zhang X, Ren S, et al. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2015, 37 (9): 1904- 1916.
doi: 10.1109/TPAMI.2015.2389824
|
|
He J, Erfani S, Ma X, et al. 2021. Alpha-IoU: a family of power intersection over union losses for bounding box regression. arXiv Preprint arXiv: 2110.13675.
|
|
Han K, Wang Y, Chen H, et al. A survey on vision transformer. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022, 45 (1): 87- 110.
|
|
Jannat F E, Willis A R. 2022. Improving classification of remotely sensed images with the Swin transformer. SoutheastCon 2022, IEEE, 611–618.
|
|
Li Y, Mao H, Girshick R, et al. 2022. Exploring plain vision transformer backbones for object detection. Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022. Proceedings, Part IX. Cham: Springer Nature Switzerland, 280−296.
|
|
Lin T Y, Goyal P, Girshick R, et al. 2017. Focal loss for dense object detection. Proceedings of the IEEE International Conference on Computer Vision, 2980−2988.
|
|
Lin T Y, Maire M, Belongie S, et al. 2014. Microsoft coco: common objects in context. European Conference on Computer Vision, Springer, 740–755.
|
|
Liu T, Ma Y, Yang W, et al. Spatial-temporal interaction learning based two-stream network for action recognition. Information Sciences, 2022, 606, 864- 876.
|
|
Liu W, Anguelov D, Erhan D, et al. Ssd: single shot multibox detector. European Conference on Computer Vision. Springer, 2016, Cham, 21- 37.
|
|
Liu Z, Tan Y, He Q, et al. Swinnet: Swin transformer drives edge-aware rgb-d and rgb-t salient object detection. IEEE Transactions on Circuits and Systems for Video Technology, 2021, 32 (7): 4486- 4497.
|
|
Liu Z, Lin Y, Cao Y, et al. 2021. Swin transformer: Hierarchical vision transformer using shifted windows. Proceedings of the IEEE/CVF International Conference on Computer Vision, 10012−10022.
|
|
Naseer M M, Ranasinghe K, Khan S H, et al. Intriguing properties of vision transformers. Advances in Neural Information Processing Systems, 2021, 34, 23296- 23308.
|
|
Khan S, Naseer M, Hayat M, et al. Transformers in vision: a survey. ACM Computing Surveys (CSUR), 2022, 54 (10s): 1- 41.
|
|
Norouzzadeh M S, Nguyen A, Kosmala M, et al. Automatically identifying, counting, and describing wild animals in camera-trap images with deep learning. Proceedings of the National Academy of Sciences, 2018, 115 (25): E5716- E5725.
|
|
Redmon J, Divvala S, Girshick R, et al. 2016. You only look once: unified, real-time object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 779−788.
|
|
Redmon J, Farhadi A. 2017. YOLO9000: better, faster, stronger. Proceedings of the IEEE Conference on Computer vision and Pattern Recognition. 7263−7271.
|
|
Redmon J, Farhadi A. 2018. Yolov3: an incremental improvement. arXiv Preprint arXiv: 1804.02767.
|
|
Ren S, He K, Girshick R, et al. 2015. Faster r-cnn: towards real-time object detection with region proposal networks. Advances in Neural Information Processing Systems, 28.
|
|
Sermanet P, Eigen D, Zhang X, et al. 2013. Overfeat: integrated recognition, localization and detection using convolutional networks. arXiv Preprint arXiv: 1312.6229.
|
|
Schneider T C, Kowalczyk R, Köhler M. 2013.Resting site selection by large herbivores―the case of European bison (Bison bonasus) in Biaowieza Primeval Forest. Mammalian Biology, 78(6): 438−445.
|
|
Villa A G, Salazar A, Vargas F. Towards automatic wild animal monitoring: identification of animal species in camera-trap images using very deep convolutional neural networks. Ecological Informatics, 2017, 41, 24- 32.
doi: 10.1016/j.ecoinf.2017.07.004
|
|
Yun S, Han D, Oh S J, et al. 2019. Cutmix: Regularization strategy to train strong classifiers with localizable features. Proceedings of the IEEE/CVF International Conference on Computer Vision, 6023−6032.
|
|
Zheng Z, Wang P, Liu W, et al. 2020. Distance-IoU loss: faster and better learning for bounding box regression. Proceedings of the AAAI Conference on Artificial Intelligence. 34(7): 12993−13000.
|