A general deep learning model for bird detection in high-resolution airborne imagery
- PMID: 35708073
- DOI: 10.1002/eap.2694
A general deep learning model for bird detection in high-resolution airborne imagery
Abstract
Advances in artificial intelligence for computer vision hold great promise for increasing the scales at which ecological systems can be studied. The distribution and behavior of individuals is central to ecology, and computer vision using deep neural networks can learn to detect individual objects in imagery. However, developing supervised models for ecological monitoring is challenging because it requires large amounts of human-labeled training data, requires advanced technical expertise and computational infrastructure, and is prone to overfitting. This limits application across space and time. One solution is developing generalized models that can be applied across species and ecosystems. Using over 250,000 annotations from 13 projects from around the world, we develop a general bird detection model that achieves over 65% recall and 50% precision on novel aerial data without any local training despite differences in species, habitat, and imaging methodology. Fine-tuning this model with only 1000 local annotations increases these values to an average of 84% recall and 69% precision by building on the general features learned from other data sources. Retraining from the general model improves local predictions even when moderately large annotation sets are available and makes model training faster and more stable. Our results demonstrate that general models for detecting broad classes of organisms using airborne imagery are achievable. These models can reduce the effort, expertise, and computational resources necessary for automating the detection of individual organisms across large scales, helping to transform the scale of data collection in ecology and the questions that can be addressed.
Keywords: airborne monitoring; bird detection; computer vision; deep learning; unoccupied aerial vehicle.
© 2022 The Ecological Society of America.
References
REFERENCES
-
- Afán, I., M. Máñez, and R. Díaz-Delgado. 2018. “Drone Monitoring of Breeding Waterbird Populations: The Case of the Glossy Ibis.” Drones 2: 42.
-
- Ahumada, J. A., E. Fegraus, T. Birch, N. Flores, R. Kays, T. G. O'Brien, J. Palmer, et al. 2020. “Wildlife Insights: A Platform to Maximize the Potential of Camera Trap and Other Passive Sensor Wildlife Data for the Planet.” Environmental Conservation 47: 1-6.
-
- Beery, S., G. Wu, V. Rathod, R. Votel, and J. Huang. 2020. “Context R-CNN: Long Term Temporal Context for per-Camera Object Detection.” 13075-13085.
-
- Beijbom, O., P. J. Edmunds, C. Roelfsema, J. Smith, D. I. Kline, B. P. Neal, M. J. Dunlap, et al. 2015. “Towards Automated Annotation of Benthic Survey Images: Variability of Human Experts and Operational Modes of Automation.” PLoS One 10: e0130312.
-
- Berger-Wolf, T. Y., D. I. Rubenstein, C. V. Stewart, J. A. Holmberg, J. Parham, S. Menon, J. Crall, J. Van Oast, E. Kiciman, and L. Joppa. 2017. “Wildbook: Crowdsourcing, Computer Vision, and Data Science for Conservation.” arXiv:1710.08880 [cs].
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources