A deep learning approach to photo–identification demonstrates high performance on two dozen cetacean species

Philip T. Patton* (Corresponding Author), Ted Cheeseman, Kenshin Abe, Taiki Yamaguchi, Walter Reade, Ken Southerland, Addison Howard, Erin M. Oleson, Jason B. Allen, Erin Ashe, Aline Athayde, Robin W. Baird, Charla Basran, Elsa Cabrera, John Calambokidis, Júlio Cardoso, Emma L. Carroll, Amina Cesario, Barbara J. Cheney, Enrico CorsiJens Currie, John W. Durban, Erin A. Falcone, Holly Fearnbach, Kiirsten Flynn, Trish Franklin, Wally Franklin, Bárbara Galletti Vernazzani, Tilen Genov, Marie Hill, David R. Johnston, Erin L. Keene, Sabre D. Mahaffy, Tamara L. McGuire, Liah McPherson, Catherine Meyer, Robert Michaud, Anastasia Miliou, Dara N. Orbach, Heidi C. Pearson, Marianne H. Rasmussen, William J. Rayment, Caroline Rinaldi, Renato Rinaldi, Salvatore Siciliano, Stephanie Stack, Beatriz Tintore, Leigh G. Torres, Jared R. Towers, Cameron Trotter, Reny Tyson Moore, Caroline R. Weir, Rebecca Wellard, Randall Wells, Kymberly M. Yano, Jochen R. Zaeschmar, Lars Bejder

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

Researchers can investigate many aspects of animal ecology through noninvasive photo–identification. Photo–identification is becoming more efficient as matching individuals between photos is increasingly automated. However, the convolutional neural network models that have facilitated this change need many training images to generalize well. As a result, they have often been developed for individual species that meet this threshold. These single-species methods might underperform, as they ignore potential similarities in identifying characteristics and the photo–identification process among species.
In this paper, we introduce a multi-species photo–identification model based on a state-of-the-art method in human facial recognition, the ArcFace classification head. Our model uses two such heads to jointly classify species and identities, allowing species to share information and parameters within the network. As a demonstration, we trained this model with 50,796 images from 39 catalogues of 24 cetacean species, evaluating its predictive performance on 21,192 test images from the same catalogues. We further evaluated its predictive performance with two external catalogues entirely composed of identities that the model did not see during training.
The model achieved a mean average precision (MAP) of 0.869 on the test set. Of these, 10 catalogues representing seven species achieved a MAP score over 0.95. For some species, there was notable variation in performance among catalogues, largely explained by variation in photo quality. Finally, the model appeared to generalize well, with the two external catalogues scoring similarly to their species' counterparts in the larger test set.
From our cetacean application, we provide a list of recommendations for potential users of this model, focusing on those with cetacean photo–identification catalogues. For example, users with high quality images of animals identified by dorsal nicks and notches should expect near optimal performance. Users can expect decreasing performance for catalogues with higher proportions of indistinct individuals or poor quality photos. Finally, we note that this model is currently freely available as code in a GitHub repository and as a graphical user interface, with additional functionality for collaborative data management, via Happywhale.com.
Original languageEnglish
Pages (from-to)2611-2625
Number of pages15
JournalMethods in Ecology and Evolution
Volume14
Issue number10
Early online date13 Jul 2023
DOIs
Publication statusPublished - Oct 2023

Bibliographical note

We thank the countless individuals who collected and/or processed the nearly 85,000 images used in this study and those who assisted, particularly those who sorted these images from the millions that did not end up in the catalogues. Additionally, we thank the other Kaggle competitors who helped develop the ideas, models and data used here, particularly those who released their datasets to the public. The graduate assistantship for Philip T. Patton was funded by the NOAA Fisheries QUEST Fellowship. This paper represents HIMB and SOEST contribution numbers 1932 and 11679, respectively. The technical support and advanced computing resources from University of Hawaii Information Technology Services—Cyberinfrastructure, funded in part by the National Science Foundation CC* awards # 2201428 and # 2232862 are gratefully acknowledged. Every photo–identification image was collected under permits according to relevant national guidelines, regulation and legislation.

Data Availability Statement

The competition data are freely available at https://www.kaggle.com/competitions/happy-whale-and-dolphin. The data and code necessary to train, validate, and test the model are available at Zenodo (Abe, 2023) and GitHub https://github.com/knshnb/kaggle-happywhale-1st-place.

Keywords

  • artificial intelligence
  • cetacean
  • computer vision
  • convolutional neural network
  • deep learning
  • dolphin
  • dorsal
  • lateral
  • machine learning
  • multi–species
  • photo–identification
  • whale

Fingerprint

Dive into the research topics of 'A deep learning approach to photo–identification demonstrates high performance on two dozen cetacean species'. Together they form a unique fingerprint.

Cite this