Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published January 2021 | Supplemental Material + Published
Journal Article Open

A deep active learning system for species identification and counting in camera trap images

Abstract

1. A typical camera trap survey may produce millions of images that require slow, expensive manual review. Consequently, critical conservation questions may be answered too slowly to support decision‐making. Recent studies demonstrated the potential for computer vision to dramatically increase efficiency in image‐based biodiversity surveys; however, the literature has focused on projects with a large set of labeled training images, and hence many projects with a smaller set of labeled images cannot benefit from existing machine learning techniques. Furthermore, even sizable projects have struggled to adopt computer vision methods because classification models overfit to specific image backgrounds (i.e., camera locations). 2. In this paper, we combine the power of machine intelligence and human intelligence via a novel active learning system to minimize the manual work required to train a computer vision model. Furthermore, we utilize object detection models and transfer learning to prevent overfitting to camera locations. To our knowledge, this is the first work to apply an active learning approach to camera trap images. 3. Our proposed scheme can match state‐of‐the‐art accuracy on a 3.2 million image dataset with as few as 14,100 manual labels, which means decreasing manual labeling effort by over 99.5%. Our trained models are also less dependent on background pixels, since they operate only on cropped regions around animals. 4. The proposed active deep learning scheme can significantly reduce the manual labor required to extract information from camera trap images. Automation of information extraction will not only benefit existing camera trap projects, but can also catalyze the deployment of larger camera trap arrays.

Additional Information

© 2020 British Ecological Society. Issue Online: 10 January 2021; Version of Record online: 20 November 2020; Accepted manuscript online: 14 October 2020; Manuscript accepted: 04 September 2020; Manuscript received: 02 February 2020. Peer Review: The peer review history for this article is available at https://publons.com/publon/10.1111/2041‐210X.13504. Data Availability Statement: The released version of the source code is available at (Norouzzadeh, 2020). Trained models and the live version of the source code are available at (https://github.com/microsoft/CameraTraps/tree/norouzzadeh‐et‐al‐2020/research/active_learning). The North American Camera Trap Images (NACTI), Snapshot Serengeti (SS) and Caltech Camera Traps (CCT) datasets are accessible in the Labeled Information Library of Alexandria: Biology & Conservation (LILA BC) digital repository (http://lila.science/datasets/).

Attached Files

Published - MethodsEcolEvol_2020-Norouzzadeh.pdf

Supplemental Material - mee313504-sup-0001-supinfo.pdf

Files

MethodsEcolEvol_2020-Norouzzadeh.pdf
Files (2.0 MB)
Name Size Download all
md5:3dac4508b880cc27b5af1d12d7eb388f
1.7 MB Preview Download
md5:e562d2ce2446d524b22b506f2025a338
237.4 kB Preview Download

Additional details

Created:
October 3, 2023
Modified:
October 24, 2023