ArticlePDF Available

Remote sensing techniques for automated marine mammals detection: a review of methods and current challenges

Taylor & Francis
PeerJ
Authors:

Abstract

Marine mammals are under pressure from multiple threats, such as global climate change, bycatch, and vessel collisions. In this context, more frequent and spatially extensive surveys for abundance and distribution studies are necessary to inform conservation efforts. Marine mammal surveys have been performed visually from land, ships, and aircraft. These methods can be costly, logistically challenging in remote locations, dangerous to researchers, and disturbing to the animals. The growing use of imagery from satellite and unoccupied aerial systems (UAS) can help address some of these challenges, complementing crewed surveys and allowing for more frequent and evenly distributed surveys, especially for remote locations. However, manual counts in satellite and UAS imagery remain time and labor intensive, but the automation of image analyses offers promising solutions. Here, we reviewed the literature for automated methods applied to detect marine mammals in satellite and UAS imagery. The performance of studies is quantitatively compared with metrics that evaluate false positives and false negatives from automated detection against manual counts of animals, which allows for a better assessment of the impact of miscounts in conservation contexts. In general, methods that relied solely on statistical differences in the spectral responses of animals and their surroundings performed worse than studies that used convolutional neural networks (CNN). Despite mixed results, CNN showed promise, and its use and evaluation should continue. Overall, while automation can reduce time and labor, more research is needed to improve the accuracy of automated counts. With the current state of knowledge, it is best to use semi-automated approaches that involve user revision of the output. These approaches currently enable the best tradeoff between time effort and detection accuracy. Based on our analysis, we identified thermal infrared UAS imagery as a future research avenue for marine mammal detection and also recommend the further exploration of object-based image analysis (OBIA). Our analysis also showed that past studies have focused on the automated detection of baleen whales and pinnipeds and that there is a gap in studies looking at toothed whales, polar bears, sirenians, and mustelids.
Submitted 11 February 2022
Accepted 13 May 2022
Published 20 June 2022
Corresponding author
Esteban N. Rodofili, erodofili@ufl.edu
Academic editor
Patricia Gandini
Additional Information and
Declarations can be found on
page 15
DOI 10.7717/peerj.13540
Copyright
2022 Rodofili et al.
Distributed under
Creative Commons CC-BY 4.0
OPEN ACCESS
Remote sensing techniques for automated
marine mammals detection: a review of
methods and current challenges
Esteban N. Rodofili1, Vincent Lecours1,2and Michelle LaRue3,4
1School of Natural Resources and Environment, University of Florida, Gainesville, FL, United States of
America
2School of Forest, Fisheries, and Geomatics Sciences, University of Florida, Gainesville, FL, United States of
America
3School of Earth and Environment, University of Canterbury, Christchurch, New Zealand
4Department of Earth and Environmental Science, University of Minnesota, Minneapolis, MN, United States
of America
ABSTRACT
Marine mammals are under pressure from multiple threats, such as global climate
change, bycatch, and vessel collisions. In this context, more frequent and spatially
extensive surveys for abundance and distribution studies are necessary to inform
conservation efforts. Marine mammal surveys have been performed visually from land,
ships, and aircraft. These methods can be costly, logistically challenging in remote
locations, dangerous to researchers, and disturbing to the animals. The growing use of
imagery from satellite and unoccupied aerial systems (UAS) can help address some of
these challenges, complementing crewed surveys and allowing for more frequent and
evenly distributed surveys, especially for remote locations. However, manual counts
in satellite and UAS imagery remain time and labor intensive, but the automation
of image analyses offers promising solutions. Here, we reviewed the literature for
automated methods applied to detect marine mammals in satellite and UAS imagery.
The performance of studies is quantitatively compared with metrics that evaluate
false positives and false negatives from automated detection against manual counts of
animals, which allows for a better assessment of the impact of miscounts in conservation
contexts. In general, methods that relied solely on statistical differences in the spectral
responses of animals and their surroundings performed worse than studies that used
convolutional neural networks (CNN). Despite mixed results, CNN showed promise,
and its use and evaluation should continue. Overall, while automation can reduce time
and labor, more research is needed to improve the accuracy of automated counts.
With the current state of knowledge, it is best to use semi-automated approaches that
involve user revision of the output. These approaches currently enable the best tradeoff
between time effort and detection accuracy. Based on our analysis, we identified thermal
infrared UAS imagery as a future research avenue for marine mammal detection and
also recommend the further exploration of object-based image analysis (OBIA). Our
analysis also showed that past studies have focused on the automated detection of baleen
whales and pinnipeds and that there is a gap in studies looking at toothed whales, polar
bears, sirenians, and mustelids.
How to cite this article Rodofili EN, Lecours V, LaRue M. 2022. Remote sensing techniques for automated marine mammals detection: a
review of methods and current challenges. PeerJ 10:e13540 http://doi.org/10.7717/peerj.13540
Subjects Conservation Biology, Ecology, Marine Biology, Zoology, Spatial and Geographic
Information Science
Keywords Accuracy metrics, Conservation surveys, Object-based image analysis, Remote sensing,
Thermal infrared
INTRODUCTION
Marine mammals currently face various anthropogenic threats such as fishery bycatch,
vessel strikes, competition for resources with commercial fisheries, noise pollution,
bioaccumulation of pathogens and toxins, harmful algal blooms, and climate change
(Merrick, Silber & De Master, 2018). In this context, surveying efforts to improve knowledge
of marine mammal species’ distributions and abundance are critical for their conservation.
Surveys can be carried out from land, aircraft or ships for cetaceans (Hiby & Hammond,
1989). Pinnipeds have been usually counted from land-based viewing stations, although
aerial photographs have also provided a viable alternative (Moore, Forney & Weller, 2018).
Boat surveys around rookeries are also commonly performed (e.g., Arias-del Razo et
al., 2016;Adame et al., 2017). However, crewed aerial and vessel platforms can bias
observations due to the animals’ reactions to them (Würsig et al., 1998;Born Riget, Dietz
& Andriashek, 1999;Dawson et al., 2004;Luksenburg & Parsons, 2009;Hashim & Jaaman,
2011). In addition, crewed aerial and boat surveys, as well as ground surveys, can be
dangerous to researchers (Sasse, 2003;Hodgson, Kelly & Peel, 2013;Gooday et al., 2018),
expensive, and logistically challenging for remote populations far away from an airstrip or
a port (Hodgson, Kelly & Peel, 2013;Gooday et al., 2018;Höschle et al., 2021).
In this context, satellite and unoccupied aerial systems (UAS) imagery have surged as
complementary tools to count individuals of various taxa of marine mammals (Sirenia,
Ursidae, Pinnipedia, and Cetacea: cf. Table 1). These two platforms share the advantage of
minimizing danger to researchers. Furthermore, the two platforms can provide still images
researchers can revisit and use to cross-validate their counts with other researchers (e.g.,
Gon¸
calves, Spitzbart & Lynch, 2020). In particular, very-high-resolution (VHR) satellite
imagery (i.e., sub-meter spatial resolution) allows monitoring marine mammals in remote
areas (LaRue, Stapleton & Anderson, 2017) without any disturbance to wildlife (LaRue
et al., 2011). As a result of a wide geographic coverage, VHR satellite images have the
potential to fill knowledge gaps on distribution, abundance, density, and population trends
of different marine mammal taxa (cf. Table 1) while supplementing data from field-based
surveys and allowing for continued monitoring when in-person surveys are not possible
(Höschle et al., 2021). A comparison between satellite and ship survey density estimates
showed encouraging results for satellite imagery (i.e., Bamford et al., 2020). On the other
hand, despite still requiring fieldwork and offering less area coverage than satellites, UAS
allow researchers more freedom in terms of survey timing than satellites and permit
coverage in regions of perennial overcast weather (e.g., Goebel et al., 2015). Adame et al.
(2017) found their UAS survey to be more accurate than a traditional boat-based survey
for the study of a California sea lion (Zalophus californianus) rookery. Furthermore, UAS
collect multispectral imagery of higher resolution than satellites, at centimeter (Johnston,
2019) and even sub-centimeter spatial resolutions (Raoult et al., 2020), which is especially
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 2/22
relevant for smaller marine mammals like mustelids (e.g., sea otters - Enhydra lutris).
Moreover, UAS can provide thermal infrared (IR) imagery at sufficient resolution to detect
marine mammals on land (e.g., Seymour et al., 2017), unlike satellite-based thermal IR,
which offers too coarse of a resolution (e.g., United States Geological Survey, 2020).
Despite their advantages, satellite and UAS platforms are not panaceas. The cost of
satellite imagery still constitutes a limiting factor for many studies (Turner et al., 2015;
Höschle et al., 2021). UAS, on the other hand, can be disturbing to animals depending
on their noise profile (Pomeroy, O’Connor & Davies, 2015). Furthermore, high-endurance
UAS recommended to cover large regions (Raoult et al., 2020) are costly, and their use can
be limited based on civil aviation regulations (Fiori et al., 2017). In such cases, satellites
remain the most cost-effective solution to study large-bodied animals over large areas
(Johnston, 2019). In addition to these issues that depend on costs, regulatory framework,
and technological advancements of these platforms, manual counts of animals in remote
sensing data are extremely labor and time-intensive for researchers over large areas (Hollings
et al., 2018;Höschle et al., 2021). For example, the analysis of 425 km2(still a relatively small
area in terms of marine mammal distribution) took 24 h to be analyzed with three replicates
(eight hours per person) for the work of Thums et al. (2018), corresponding to 1.13 min per
km2. Furthermore, Cubaynes et al. (2019) reported three hours and 20 min to scan 100 km2
at a 1:1,500 m. scale (2 min per km2), and Charry et al. (2021) reported an even slower rate
of approximately 2.5 min per km2(at a 1:536 scale). While satellite images have allowed
continental-scale abundance and distribution studies in remote locations (e.g.,LaRue et
al., 2019;LaRue et al., 2020;LaRue et al., 2021), that work required crowdsourcing: LaRue
et al. (2020) enlisted more than 325,000 volunters to analyze 268,611 km2of images while
a previous effort (LaRue et al., 2019) used more than 5,000 volunteers for the Ross Sea
region in Antarctica. These broad-scale studies remain scarce for marine mammals: most
work based on satellite images is performed over smaller areas such as individual islands
(e.g., LaRue et al., 2015;LaRue & Stapleton, 2018).
Automation of image analysis reduces the level of effort and time needed, especially
for large-scale projects (Hollings et al., 2018;Charry et al., 2021;Höschle et al., 2021). Even
semi-automation in Thums et al. (2018) shortened 24 person-hours to 34 min to run the
algorithm and an additional 20 min to revise sub-sections with whale detections and discard
those that were false (a reduction of more than 96% of the analysis time). Automation could
also help with errors caused by human fatigue from reviewing imagery (Hodgson, Kelly &
Peel, 2013). As such, automation could unleash the full potential of VHR satellite images
to cover more extensive study areas, which is ultimately a capability that distinguishes the
use of satellite imagery from crewed surveys. Automating animal detection in UAS imagery
has also been suggested to provide better survey data (Hodgson, Peel & Kelly, 2017). Just as
for satellite imagery, automation would also reduce the labor and time necessary to count
animals from UAS imagery (Linchant et al., 2015). Technologies for automated count of
animals have emerged for areas difficult to access, in which ground counts are difficult, or
for animals that are at low densities over large areas (Hollings et al., 2018). Marine mammals
generally fit all these conditions, given their aquatic habitats and extended ranges, making
them suitable candidates for the application of automated counts. However, automated
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 3/22
Table 1 Marine mammal automated detection studies using satellite or UAS imagery. Imagery type is based on testing imagery, although in sev-
eral studies the training imagery was from the same source. Automated studies results were assessed with Eqs. (3),(4) and (5). Missed animals refer
to Eq. (3), false animals refer to Eq. (4), and total deviation refers to Eq. (5). For automation results assessment calculations, see Eqs. (3),(4) and (5),
and Table S1.
Study Taxa Platform; altitude
(if UAS); imagery
type (spatial resolution)
Automated
method
Automation results
assessment
Mejias Alvarez et
al. (2013)
Dugongs (Dugong
dugon) (Sirenia)
UAS (ScanEagle) with Nikon
12 MP. digital SLR camera
and 50 mm. lens and polar-
izing filter; 500/750/1000 ft.;
RGB (ground sampling reso-
lution not found)
Morphological based detec-
tion, segmentation, shape
profiling on saturation chan-
nel
Not enough data was found to cal-
culate Eqs. (3),(4) and (5)
Maire et al. (2013) Dugongs (Dugong
dugon) (Sirenia)
UAS (not specified) with
Nikon 12 MP. digital
SLR camera and 50 mm.
lens and polarizing filter;
500/750/1000 ft.; RGB
(ground sampling resolution
not found)
Color and morphological fil-
ters, segmentation and shape
analysis
Not enough data was found to cal-
culate Eqs. (3),(4) and (5)
Fretwell, Staniland
& Forcada (2014)
Southern right
whales (Eubalena
australis) (Mysticeti)
Satellite (WorldView-2);
panchromatic (0.5 m.) and
multispectral (2 m. but pan-
sharpened)
Unsupervised classification,
supervised classification and
histogram thresholding
Supervised classification:
no meaningful results
Unsupervised classification
Kmeans: 41.76%
(missed), 53.85% (false),
95.6% (total deviation)
Histogram thresholding Band
5: 15.38% (missed), 26.37%
(false), 41.76% (total deviation)
(Best variants within methods
chosen based on Eq. (5) for total
signals, not by probable, possible
and band 5 manual detections)
LaRue et al. (2015) Polar bears (Ursus
maritimus) (Ursi-
dae)
Satellite (WorldView-2/
QuickBird)**; panchromatic
(0.5-0.65 m.)*
Supervised classification and
image differencing
Supervised classifi-
cation: unsuccessful.
Image differencing: not enough
data was found to calculate Eqs.
(3),(4) and (5).
Seymour et al.
(2017)
Grey seals
(Halichoerus
grypus) (Pinnipedia)
UAS (senseFly eBee) with 12
MP. RGB Canon S110 cam-
era and 640 ×512-pixel ther-
mal IR senseFly LLC, Ther-
momapper camera; altitude
not found; RGB (3 cm.) and
thermal IR (8 cm.)
ArcGis Model based on tem-
perature, size and shape
Pups: Saddle Island (simple):
6.45% (missed), 13.55%
(false), 20% (total deviation)
Saddle Island (Complex):
7.04% (missed), 13.9% (false),
20.94% (total deviation)
Adults: Saddle Island (simple):
1% (missed), 23.38% (false),
24.38% (total deviation)
Saddle Island (Complex): 0.98%
(missed), 49.02% (false), 50%
(total deviation) (only models
from prediction site)
(continued on next page)
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 4/22
Table 1 (continued)
Study Taxa Platform; altitude
(if UAS); imagery
type (spatial resolution)
Automated
method
Automation results
assessment
Thums et al. (2018) Humpback whales
(Megaptera no-
vaeangliae) (Mys-
ticeti)
Satellite (WorldView-2/
WorldView-3); panchromatic
(0.4 m.) and multispectral
(but pansharpened) /
panchromatic (0.31 m.)
Unsupervised classification
and supervised classification
(in Worldview 2 imagery)
and semi-automated algo-
rithm based on shape (in
Worldview-3 imagery)
Unsupervised classification and
supervised classification: not
enough data found for Eqs. (3),
(4) and (5). Shape algorithm:
0% (missed), 205.8% (false),
205.8% (total deviation) (calves
and mothers together, all images
combined)
Gray et al. (2019) Blue whales
(Balaenoptera
musculus) and
humpback whales
(Megaptera
novaeangliae)
and Antarctic
minke whales
(Balaenoptera
bonaerensis)
(Mysticeti)
UAS (FreeFly Alta 6/LemHex-
44) with Sony a5100 cam-
era with 50 mm. Focal length
lens, 23.5 ×15.6 mm. Sen-
sor size and 6000 ×4000 pixel
resolution; 30–80 m.; RGB
(see equation in study for
ground sampling distance)
CNN (deep learning) 0% (missed), 1.72% (false), 1.72%
(total deviation) (only for whale
recognition, not species)
Borowicz et al.
(2019)
Southern right
whales (Eubalaena
australis) and
humpback whales
(Megaptera
novaeangliae)
(Mysticeti)
Satellite (WorldView-3); mul-
tispectral (1.24 m. but pan-
sharpened to 0.31 m.)
CNN (deep learning) 0% (missed), 271.88% (false),
271.88% (total deviation) (best
model chosen by authors)
Guirado et al.
(2019)
Various Google Earth Imagery
(included USGS
aerial/WorldView-
3/QuickBird-2/GeoEye-
1/SPOT-6/WorldView-
2); RGB (0.15 m./0.31
m. panchromatic–1.24
m. multispectral/0.61
m. panchromatic–2.5
m. multispectral/0.46
m. panchromatic–1.84
m. multispectral/1.5
m. panchromatic–6 m.
multispectral/0.46 m.
panchromatic–1.84 m.
multispectral)
CNN (deep learning) Detection CNN: 20.59% (missed),
7.35% (false), 27.94% (total de-
viation). Count CNN: 11.43%
(missed), 4.29% (false), 15.71%
(total deviation) (all locations to-
tals)
(continued on next page)
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 5/22
Table 1 (continued)
Study Taxa Platform; altitude
(if UAS); imagery
type (spatial resolution)
Automated
method
Automation results
assessment
Cubaynes (2019) Southern right
whales (Eubalaena
australis) (Mysticeti)
Satellite (GeoEye-1); mul-
tispectral (1.65 m. but pan-
sharpened to 0.41 m.)
Unsupervised classification,
supervised classification,
thresholding and OBIA
Unsupervised classification
Isodata: 56.82% (missed),
140.91% (false), 197.73%
(total deviation).
Supervised classification
(maximum likelihood): 9.09%
(missed), 3.41% (false),
12.5% (total deviation).
Thresholding (only NIR1):
34.09% (missed), 194.32%
(false), 228.41% (total deviation).
OBIA: 27.27% (missed), 453.41%
(false), 480.68% (total deviation).
(for total whales, not by definite,
probable and possible)
Gon¸
calves,
Spitzbart &
Lynch (2020)
Crabeater
seals (Lobodon
carcinophaga),
Weddell seals
(Leptonychotes
weddellii), leopard
seals (Hydrurga
leptonyx) and Ross
seals (Omnatophoca
rossii) (Pinnipedia)
Satellite (WorldView-3);
panchromatic (0.3 m.)
CNN (deep learning) SealNet: 69.78% (missed), 51.71%
(false), 121.49% (total deviation)
(total for all scenes) (only model
with best F1 for testing)
Zinglersen et al.
(2020)
Walruses (Odobenus
rosmarus) (Pinni-
pedia)
Satellite (Pléiades
1 A/B/WorldView-
2/WorldView-3);
multispectral (2 m. but
pansharpened to 0.5 m./1.84
m. but pansharpened to 0.46
m./1.24 m. but pansharpened
to 0.31 m.)
OBIA Not enough data found for Eqs.
(3),(4) and (5).
Dujon et al. (2021) Australian fur seal
(Arctocephalus pusil-
lus) (Pinnipedia)
UAS (DJI Phantom 4
ProfessionalTM V2) with
built-in camera; 35 m.; RGB
(ground sampling resolution
not found)
CNN (Deep learning) Not enough data found for Eqs.
(3),(4) and (5).
Notes.
Information obtained from Hollings et al. (2018)
∗∗Information obtained from Cubaynes (2019)
detection in UAS imagery has been more often applied to terrestrial mammals and birds
than to marine mammals (Corcoran et al., 2021), and Hollings et al. (2018) report a similar
pattern for automated analysis of remote sensing imagery in general.
Automated analysis of remote sensing imagery can be divided into pixel or object-based
methods, depending on the unit of analysis (Blaschke et al., 2014;Wang, Shao & Yue,
2019). While pixel-based methods (e.g., unsupervised or supervised classification and
thresholding) classify individual pixels into animal or background classes based on spectral
information or texture, object-based methods classify groups of homogeneous pixels
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 6/22
into the different classes, which allows for the use of other variables, such as size and
shape (Blaschke et al., 2014;Wang, Shao & Yue, 2019). This contribution reviews the use
of different methods of automated analysis of VHR satellite and UAS imagery for marine
mammal detection, which is a current gap as previously-published reviews are dedicated to
wildlife in general, either to their automated detection in UAS (i.e.,Corcoran et al., 2021)
or manual and automatic detection in UAS, satellite and aerial imagery (i.e., Hollings et al.,
2018;Wang, Shao & Yue, 2019). The detection of marine mammals in their natural habitats
presents different challenges that warrant their own review. The analysis we present also
adds to contributions that focused on using satellite (LaRue, Stapleton & Anderson, 2017;
Höschle et al., 2021) and UAS (Koski, Abgrall & Yazvenko, 2010;Smith et al., 2016;Fiori et
al., 2017;Johnston, 2019;Schofield et al., 2019;Raoult et al., 2020) imagery to study marine
wildlife separately. This review aims to inform the multidisciplinary community of remote
sensing experts, biologists, ecologists, and managers working on marine mammal research
and conservation as to which automated methods and imagery type show more promise,
but also which taxa have been overlooked, practical alternatives to ensure time-saving
and accurate counts, and steps necessary to progress from individual studies to the
regional-scale initiatives that conservation needs. Our analysis is also summarized in Table
1, which improves upon the work of Hollings et al. (2018) and Cubaynes (2019) by the
application of uniform metrics for evaluating detection accuracy that allow comparing the
results of existing studies directly and for a more straightforward assessment of over- and
undercounts in a conservation context.
SURVEY METHODOLOGY
The search engine of the University of Florida Libraries (https://uflib.ufl.edu/find/) Primo
database was used repeatedly between August 2019 to January 2022 to identify peer-
reviewed literature, conference proceedings, grey literature, and theses and dissertations
about combinations of keywords such as ‘‘marine mammals’’, ‘‘remote sensing’’, ‘‘UAV’’,
‘‘UAS’’, ‘‘satellite’’ and ‘‘images’’. The Primo database searches across the majority of
the authoritative electronic and print resources subscribed to and purchased by the
University of Florida Libraries (UF George A. Smathers Libraries, 2022). The literature cited
in these studies was then explored further if relevant to image analysis for marine mammal
detection, either through Primo or Google Scholar. We excluded studies of detection of
marine mammal groups (e.g., Burn & Cody, 2005;Fischbach & Douglas, 2021), traces and
holes (e.g., Platonov, Mordvintsev & Rozhnov, 2013), carcasses (e.g., Fretwell et al., 2019;
Clarke et al., 2021), and works that compared automated detection in separate datasets of
different spatial resolutions (e.g., Fischbach & Douglas, 2021;Corrêa et al., 2022).
CURRENT TRENDS IN MARINE MAMMAL AUTOMATED
DETECTION
We found 13 studies on automated detection of marine mammals ranging from 2013 to
2021, of which five used UAS imagery and eight used satellite imagery (Table 1). While
many different automation approaches have been tested, most studies used pixel-based
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 7/22
approaches (as opposed to object-based workflows). The early attempts in pixel-based
automation used spectral information through unsupervised classifications, supervised
classifications, histogram thresholding, and image differencing (Table 1), for example
for the classification of Southern right whales (Eubalaena australis) (Fretwell, Staniland
& Forcada, 2014) and polar bears (Ursus maritimus) (LaRue et al., 2015). Histogram
thresholding consists of maximizing the class of interest’s signal to reduce the amount
of noise (Fretwell, Staniland & Forcada, 2014). Automated image differencing is a process
to detect differences in pixels between satellite images collected at different times (Singh,
1989;Lu et al., 2004;LaRue et al., 2015). These methods proved not to be very effective in
providing accurate counts (cf. Table 1). This aligns with recommendations from Cubaynes et
al. (2019) to adopt approaches that can use more information (e.g., topological, geometrical,
or textural information) than only spectral responses, such as deep learning or object-based
image analysis (OBIA).
Convolutional neural networks (CNN), a deep learning approach, have been increasingly
used to analyze satellite and UAS images of whale hotspots (Borowicz et al., 2019;Guirado
et al., 2019;Gray et al., 2019) and pinnipeds (Gon¸
calves, Spitzbart & Lynch, 2020;Dujon et
al., 2021) (Table 1). CNN has had mixed results so far, with relative success when applied
to satellite (Guirado et al., 2019) and UAS (Gray et al., 2019) images of whales. Of note,
CNN proved able to distinguish between different species in Gray et al. (2019), something
which had not yet been achieved from satellite images (Höschle et al., 2021). However,
not all CNN applications for whale detection were successful (Borowicz et al., 2019), and
studies that applied CNN for the detection of pinnipeds in satellite (Gon¸
calves, Spitzbart &
Lynch, 2020) and UAS (Dujon et al., 2021) imagery did not show high accuracies. Gon¸
calves,
Spitzbart & Lynch (2020) reported the highest precision and recall among different scenes
for their SealNet model as 0.519 and 0.377, respectively, in their crabeater seals (Lobodon
carcinophaga), Weddell seals (Leptonychotes weddellii), leopard seals (Hydrurga leptonyx)
and Ross seals (Omnatophoca rossii) study. Dujon et al. (2021) reported a 0.27 precision
for Australian fur seal (Arctocephalus pusillus) detection. Borowicz et al. (2019) suggested
including confounding classes such as ships, rocks, and land to minimize false positives
and improve CNN accuracy. This worked relatively well for Guirado et al. (2019) but did
not provide more accurate counts for Gon¸
calves, Spitzbart & Lynch (2020). Another way to
improve detection accuracy was the application of a two-step CNN that first detects whale
presence in individual tiles of an image collection and then locates and counts whales in
tiles that had them (Guirado et al., 2019). A valuable insight the authors offer about their
two-step approach is that they find better results than with one detection model alone, as
a result of the first CNN filtering false positives and therefore allowing the second CNN to
count whales more accurately.
CNN studies have also sparked a discussion on how marine mammal behavior impacts
detection abilities. For example, Dujon et al. (2021) developed a CNN capable of detecting,
as mentioned before, Australian fur seals (Arctocephalus pusillus), but also loggerhead sea
turtles (Caretta caretta), and Australian gannets (Morus serrator) in UAS imagery from
separate sites. The authors reference the aggregation of pups in creches as one of the factors
that made seal detection difficult compared to that of the other taxa analyzed, up to the
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 8/22
point that the CNN was not able to distinguish between multiple aggregated pups. Whales
can also offer particular detection challenges based on their behavior. For instance, Guirado
et al. (2019) concluded that more than 90% of true positives of their detection CNN were
whales in blowing, breaching, peduncle emerged, or logging behavior, whereas 33% of
false negatives were submerged whales—which happened to be the most frequent behavior
observed. The remaining 66% of false negatives were spy-hopping whales (Guirado et al.,
2019). Cubaynes et al. (2019) suggested that some species with less acrobatic behavior were
easier to detect. This is not that clear in Guirado et al.’s (2019) findings, as some whale
behaviors which could be characterized as acrobatic, i.e., with their peduncle emerged and
breaching, were part of most of the true positive cases along with blowing and logging.
However, they recognized that the spyhopping behavior compromised the CNN’s ability
to detect whales.
In terms of detectability and accuracy evaluation, while Guirado et al. (2019) and
Cubaynes et al. (2019) discussed the influence of behaviors displayed on the surface or
submerged—but still visible—on the detectability of animals, there remains the issue of
animals submerged to the point of not being visible. In general, studies that detect marine
mammals automatically in imagery evaluate the accuracy of their automation methods
by comparing the animals captured by the computer with the ones counted manually.
This approach is based on obtaining a count of the detectable animals, i.e., only those
visible in the imagery. This is a common limitation of visual surveys from land, ships,
aircraft, or using UAS or satellite imagery, and especially in marine mammal taxa that are
permanently in the water (i.e., sirenians and cetaceans). While, as mentioned before, UAS
and satellite imagery have the advantage of allowing surveyors to go over them many times
and cross-validate their counts with other researchers—helping to minimize perception
bias compared to field surveys from land, ships or aircraft—they are still visual methods not
able to detect animals submerged to the point of not being visible. Consequently, it is worth
noting that while this review is focused on the automated detection of marine mammals
in imagery to obtain animal counts as a first step, further steps are needed after the counts
to calculate abundance or density estimates. For example, there is a requirement to make
adjustements for submerged animals not showing in the imagery—animals available in the
study area but not at the surface (not detectable), as Bamford et al. (2020) did for whale
densities calculated from satellite imagery. These adjustments could involve additional
information obtained through tracking devices, such as surfacing and submersion times
(e.g., Bamford et al., 2020).
The general approach presented above—comparing objects detected automatically with
user observations assumed to correspond to all animals detectable in the image—aligns with
evaluation methods commonly used in remote sensing. When adopting this approach,
false negatives correspond to detectable animals missed by the computer. In contrast,
false positives refer to anything that is not the detection target, such as water or rocks,
and is classified wrongly as a marine mammal by the computer. False negatives thus
contribute to an undercount of marine mammals, and false positives contribute to an
overcount. However, there is currently a lack of consistency in accuracy metrics to evaluate
automated methods across animal detection studies (Hollings et al., 2018). Furthermore,
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 9/22
when comparing automated detections with manual counts to evaluate detection accuracy,
some studies defined the F1 index metric differently. For example, Borowicz et al. (2019)
used Eq. (1) and Gon¸
calves, Spitzbart & Lynch (2020) used Eq. (2). This can be highly
problematic when comparing different approaches.
F1=2precisionrecall
precision+recall .(1)
Equation 1: F1 index as defined by Borowicz et al. (2019).
F1=precisionrecall.(2)
Equation 2: F1 index as defined by Gon¸
calves, Spitzbart & Lynch (2020).
FUTURE DIRECTIONS IN MARINE MAMMAL AUTOMATED
DETECTION
Object-based image analysis (OBIA)
OBIA has been less explored for marine mammal detection compared to pixel-based
methods. While both pixel- and object-based approaches may consider more variables
than spectral information (e.g., texture) (Cubaynes et al., 2019), they are fundamentally
different in their units of analysis (Blaschke et al., 2014). OBIA first segments the imagery,
dividing it into spatially continuous objects whose internal heterogeneity is less than the
heterogeneity of their neighbors (Blaschke et al., 2014). This produces scale-dependent,
potentially more meaningful objects, or segments, made of relatively homogeneous pixels.
The classification algorithm selected is then applied to the objects rather than the individual
pixels making them up. In addition, OBIA provides more information than only the spectral
response to inform the classification: for instance, information about the objects themselves,
such as size, shape, relative or absolute location, boundary conditions, and topological
relationships, can be integrated as features within the classification, something that cannot
be done in pixel-based classifications (Blaschke et al., 2014). In theory, it would make sense
that marine mammals be detected as homogeneous objects in remote sensing imagery.
Some studies have used object features, such as shape (i.e., Seymour et al., 2017;Thums
et al., 2018), and a study used highlighting of elliptical features, segmentation, and feature
extraction, such as length and area, from blobs (i.e., Mejias Alvarez et al., 2013). Another
work by Maire et al. (2013) used shape from blobs obtained after segmentation with a
prior determination of regions of interest using color and morphological filters. Results
have not been conclusive: while the work of Seymour et al. (2017) proved successful in
detecting pinnipeds in thermal IR imagery, Thums et al. (2018),Mejias Alvarez et al. (2013)
and Maire et al. (2013) did not get as good results in detecting a baleen whale species and a
Sirenia species, respectively (cf. Table 1). Furthermore, OBIA was tested with mixed results.
For Southern right whale (Eubalaena australis), Cubaynes (2019) had a moderate amount
of missed (27.27%) animals, but a considerably high number of false positives (453.41%)
(Table 1). Zinglersen et al. (2020) obtained a good object definition for walruses (Odobenus
rosmarus), but this was dependent on site heterogeneity (i.e., the relative composition of
sand and snow impacted the analysis).
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 10/22
OBIA has not been used extensively to detect other animals either, leaving pixel-based
studies as predominant. While OBIA has been suggested to perform better than pixel-based
analyses in land classification, that is still not clear in the examples of OBIA for counting
animals (see Hollings et al., 2018). It is also worth noting that OBIA can be computationally
time-intensive (Hollings et al., 2018). Furthermore, there is a lack of an obvious method to
measure segmentation accuracy and sometimes real-life objects in the image do not match
the segments obtained, making it difficult to assess classification accuracy (Ye, Pontius &
Rakshit, 2018).
Thermal infrared (IR) imagery
Another resource worth exploring in the future for marine mammal automated image
analysis is the use of thermal IR. The combined use of shape, size, and temperature
captured from thermal IR allowed Seymour et al. (2017) to distinguish pinniped pups from
adults and count animals in aggregations, which was a recognized challenge for this taxon.
Hollings et al. (2018) argued that a high contrast between animals and their surroundings
is particularly important for automation methods, which is consistent with Cubaynes
et al. (2019) findings. Based on these arguments, thermal IR UAS imagery is promising
for marine mammal automated detection as it should assist with distinguishing marine
mammals from confounding objects or like-colored surfaces. While Hollings et al. (2018)
found that automation attempts to count animals have shown reasonably high accuracy
in areas that are small relative to a species’ range and/or homogenous environments such
as ice, the accuracy assessment metrics in Table 1 for studies that match this description
suggest this may not be necessarily the case for marine mammal detection (see Fretwell,
Staniland & Forcada, 2014;Borowicz et al., 2019;Gon¸
calves, Spitzbart & Lynch, 2020 in
Table 1). For these cases, thermal IR may be a tool worth exploring to increase the contrast
between the animals and the background. However, thermal IR has only been used for
automated detection of pinnipeds over land (i.e., Seymour et al., 2017), and as such this
recommendation can only be supported by research results in such settings. The potential
of automated analyses on UAS thermal IR data for taxa that are in the water at all times
(i.e., cetaceans and sireneans) remains a knowledge gap at this time.
Evaluation metrics
Here we assess the number of false negatives in the different studies of Table 1 using the
miss rate, or false negative rate (FNR; Eq. (3)). We consider that this metric provides
a straightforward assessment of false negatives—the undercount—in the context of
conservation and management, as its denominator is the total number of animals that
could have been detected (using the manual count as a reference). In order to have a metric
that assesses false positives—the overcount—over the total number of animals that could
have been detected too, we used a false positive over detectable rate (FPDR; Eq. (4)). In the
context of a remote sensing application to inform conservation efforts, we considered it
may be helpful to measure the undercount (all detectable animals missed by the computer)
or the overcount (all false computer detections) against the actual animals detectable in
the image, for a faster and direct assessment of how reliable an automated method would
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 11/22
be if adopted in conservation planning, and what the impact of its errors would be over
counts later used in abundance estimations.
(1recall)×100 =FN
TP +FN ×100 =FN
Detectable animals ×100.(3)
False negative rate (Equation 3): ratio between the number of animals missed by an
automated method (numerator) and the number of animals detectable in the image (the
animals counted manually) (denominator), presented as a percentage. FN: false negatives.
TP: true positives.
FP
TP +FN ×100 =FP
Detectable animals ×100.(4)
False positive over detectable rate (Eq. (4)): ratio between the number of objects wrongly
classified as animals by an automated method (numerator) and the number of animals
detectable in the image (the animals counted manually) (denominator), presented as a
percentage. FP: false positives. FN: false negatives. TP: true positives.
To summarize these two metrics into one, in the same way the F1 index integrates
precision and recall, we suggest the automated count deviation (Eq. (5)). The automated
count deviation (Eq. (5)) adds false negatives to false positives and measures them against
the detectable animals too, to help in conservation and management, as neither false
positives nor negatives are desirable in that context. The FNR (Eq. (3)), FPDR (Eq. (4))
and automated count deviation (Eq. (5)) can also be used in combination with recall,
precision and the F1 index for a more comprehensive assessment of performance.
Furthermore, we note that even though manual counts are taken as a reference of
the animals detectable in the image, this is not an infallible strategy. The effects of
environmental conditions, UAS flight-related variables, and the angle of imagery capture
on manual detection certainty of different species should continue to be investigated
(see Aniceto et al., 2018;Subhan et al., 2019). Furthermore, the subclassification of manual
detections into probable or possible (e.g., Fretwell, Staniland & Forcada, 2014) together
with cross-validated manual counts among researchers (e.g., Gon¸
calves, Spitzbart & Lynch,
2020) can help in attaining the best manual detection reference possible.
FN +FP
TP +FN ×100 =FN +FP
Detectable animals ×100.(5)
Automated count deviation (Eq. (5)): ratio between the sum of animals missed and
the objects wrongly classified as animals by an automated method (numerator) and the
animals detectable in the image (the animals counted manually) (denominator), presented
as a percentage. FN: false negatives. FP: false positives. TP: true positives.
DISCUSSION
The current threats marine mammals face justify the need for their populations to be
monitored through spatially extensive and frequent surveys for abundance and distribution
studies. Satellite and UAS imagery can help address these challenges and complement
crewed surveys. Satellite images can provide extensive coverage, enabling even continental-
scale counts to use for abundance and distribution estimations in remote locations (e.g.,
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 12/22
LaRue et al., 2019;LaRue et al., 2020;LaRue et al., 2021), making sampling efforts more
globally even. Although UAS provide less spatial coverage than satellite images, they
can allow more control over survey timing and surveying in overcast conditions, as
well as higher spatial resolution, which can be of use for smaller marine mammals like
sea otters (Enhydra lutris) or dolphins of the Cephalorhynchus genus. Nevertheless, the
manual analysis of satellite and UAS imagery is time- and effort-intensive, for which
automation can help (Linchant et al., 2015;Hollings et al., 2018;Charry et al., 2021;Höschle
et al., 2021). While Thums et al. (2018) provided a comparative measure of time saved by
their workflow, we recommend future automated marine mammal studies do the same,
i.e., that they evaluate how efficient different automated methods are in saving time and
effort compared to a manual approach.
We also acknowledge the current challenges of automated approaches in matching the
animals present in the imagery. In the studies we analyzed, automation efforts thus far
have shown considerable deviation of their detections over the animals manually counted,
in particular when it came to false positives (cf. Table 1). This is especially concerning in
a conservation assessment context, as such an error could lead to the lifting of protection
measures when they should stay in place, and therefore calls for future research to improve
detection methods. Until automated methods are better, studies on marine mammal
detection should compromise with semi-automated approaches that involve a user review
(Höschle et al., 2021). However, one positive aspect is that even if a manual review is needed,
a semi-automated method that yields image sub-sections or thumbnails that need to be
revised manually afterwards can still shorten the overall revision time, as we have seen
from Thums et al. (2018).Dujon et al. (2021) also recommend semi-automation in the
sense of leaving full automation for situations of animals that show optimal conditions for
such analysis, and doing manual detections in more challenging settings. Furthermore, the
authors recognized that while their CNN could be used in different species even within a
single site, a sole algorithm would not be able to capture all components of different species
optimally as a result of morphology, spacing behavior, and habitat (Dujon et al., 2021).
Here, we have reviewed the automated approaches for detecting marine mammals
through satellite and UAS images. We observed that the automated methods reliant
only on spectral information did not show high-performance (e.g., Fretwell, Staniland
& Forcada, 2014;LaRue et al., 2015), and spectral contrast between baleen whales and
their surroundings were found to be insufficient for automation, highlighting the need
for other methods that can accommodate more information, such as deep learning and
OBIA (Cubaynes et al., 2019). Deep learning showed high performance in some studies
(i.e., Guirado et al., 2019;Gray et al., 2019) and even allowed for distinguishing between
whale species (i.e., Gray et al., 2019), although this has not yet been achieved in satellite
images (Höschle et al., 2021). Nevertheless, not all deep learning attempts were successful,
and some attempts required hundreds (e.g., Gray et al., 2019) or thousands (e.g., Guirado
et al., 2019) of images for training. An extra complication is that labeled satellite images
are currently limited, highlighting the need for shared training datasets (Höschle et al.,
2021). Overall, CNN shows promise, and future research should continue to improve on
the studies already conducted. For instance, CNNs only successful attempts thus far have
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 13/22
been in whales—although the contrary is not true—for which further studies are due to
see if the method’s results in pinnipeds can be improved, and to evaluate the method in
other smaller marine mammal taxa. The number of studies using CNN is still low, and
more studies need to be undertaken to identify which components of workflows improve
accuracy. For example, the use of more confounding classes seemed useful for Guirado et
al. (2019) but was not a guarantee of success for Gon¸
calves, Spitzbart & Lynch (2020).
We agree with Hollings et al. (2018) in that less costly images and software are necessary
to drive the field forward. The use of free platforms for image analysis, such as Google
Earth Engine, could provide a solution by also providing greater computational power
to researchers. We recommend the development of semi-automated methods that can
be applied to the simultaneous detection of multiple species sharing a common habitat.
Therefore, free, regional, or even continental initiatives that combine training data for
recognition of different species, through the cooperation among different research groups,
government agencies or nonprofit organizations, are what is now necessary to provide a
useful tool for conservation managers to obtain counts for abundances and distribution
studies. Moreover, with readily available and accessible satellite image coverage, Abileah
(2001) and Cubaynes et al. (2019) recommendations to use satellite imagery to study
migrations could be finally implemented at a large scale, and object-based methods
could provide exclusive applications, such as the detection of swimming directions of
different individuals. This could complement the use of tagging devices and lead to
mapping migration routes for policy measures aiming to reduce collisions and bycatch.
Furthermore, accessible satellite image coverage combined with the speed of automation
could help improve distribution estimations, discover unknown feeding and calving
grounds, and keep track of distribution shifts due to climate change. Moreover, it is
worth considering baleen whales have been the focus of most automated studies, and
although some automated studies were applied to pinnipeds, sirenians, and polar bears,
more studies are needed to evaluate different platforms, automation methods, or spectral
bands for these other taxa. Automated studies on toothed whales and mustelids are also
due. Here we focused on automated detection of living marine mammal individuals in
VHR satellite and UAS imagery, but it is noteworthy that other promising satellite and
UAS image analysis applications (automated and manual) exist and focus for example on
detecting marine mammal groups (e.g., Burn & Cody, 2005;Fischbach & Douglas, 2021),
traces and holes (e.g., Platonov, Mordvintsev & Rozhnov, 2013), carcasses (e.g., Fretwell et
al., 2019;Clarke et al., 2021), and on comparing detection success from imagery of different
spatial resolutions (e.g., Fischbach & Douglas, 2021;Corrêa et al., 2022).
Regarding new promising directions to explore, OBIA has been recommended for
automated analysis (Cubaynes et al., 2019), but thus far pixel-based methods have been
more thoroughly explored and the OBIA studies or the studies using object features have
shown mixed results (e.g., Mejias Alvarez et al., 2013;Maire et al., 2013;Seymour et al.,
2017;Thums et al., 2018;Cubaynes, 2019;Zinglersen et al., 2020). Moreover, UAS offer
data at sufficient resolution in the thermal IR, which has proven useful in distinguishing
between pinniped pups and adults and counting individuals in aggregations (i.e., Seymour
et al., 2017). The use of multispectral imagery has been prevalent among studies thus far,
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 14/22
and we recommend further exploring the use of thermal IR as a solution to distinguish
animals from similarly-colored surroundings, an issue spanning different taxa and surfaces
(e.g., LaRue et al., 2015;Cubaynes et al., 2019;LaRue et al., 2020), or even to study marine
mammals nighttime behavior, or during the polar night (Boehme et al., 2016). Furthermore,
we decided to use the FNR (Eq. (3)), the FPDR (Eq. (4)) and the automated count deviation
(Eq. (5)) proposed as a more straightforward means of assessing how reliable an automated
method would be if adopted in conservation planning, and what the impact of its errors
would be over counts later used in abundance estimations.
Marine mammal abundance and distribution studies can be well served from a
combination of different surveying methods. For instance, UAS and satellite images,
airplane and vessel surveys, to detect animals in the surface; passive acoustic monitoring,
to detect submerged animals’ vocalizations (e.g., Davis et al., 2017;Stanistreet et al., 2018);
and satellite tracking, to obtain detailed movement patterns (e.g., Horton et al., 2017;
Horton et al., 2020). In the last decade, studies using satellite and UAS imagery for marine
mammal surveys have grown considerably in number, but lag behind the number of
automated studies on other taxa. This may occur because of the logistical complexity
of conducting UAS surveys at sea as opposed to on land (e.g., launching and landing
sites, weather conditions), as well as the relative scarcity of archival satellite imagery
over sea compared to terrestrial habitats, leaving researchers only with more expensive
tasking options. Automated methods have the potential to complement airplane and vessel
marine mammal surveys, helping overcome old limitations, such as higher coverage in
remote regions. However, to do that, satellite imagery needs to become more available and
affordable further away from the coast, and high-endurance UAS need to become more
affordable too, while their regulation needs to be flexibilized. In this context, joint initiatives
for further research are necessary to not only to exchange imagery, but also to develop
automated methods that both save researchers time and provide accurate counts, while
being versatile enough to work in different settings and distinguish coexistent species.
ACKNOWLEDGEMENTS
Authors acknowledge Ng¯
ai T¯
u¯
ahuriri as the University of Canterbury’s mana whenua
partner, and the Potano (Timucua) and Seminole Tribes, on whose lands the University of
Florida is located. Thanks are also due to the three reviewers and the handling editor for
their constructive comments that significantly improved earlier versions of this manuscript.
ADDITIONAL INFORMATION AND DECLARATIONS
Funding
Esteban N. Rodofili is funded jointly by the School of Natural Resources and Environment
and the School of Forest, Fisheries, and Geomatics Sciences (College of Agricultural and
Life Sciences) of the University of Florida. Funds allocated to Vincent Lecours by the
University of Florida Senior Vice President for Agriculture and Natural Resources were
also used to support this work. Support for Michelle LaRue was provided by the School
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 15/22
of Earth and Environment, University of Canterbury. There was no additional external
funding received for this study. The funders had no role in study design, data collection
and analysis, decision to publish, or preparation of the manuscript.
Grant Disclosures
The following grant information was disclosed by the authors:
School of Natural Resources and Environment.
School of Forest, Fisheries, and Geomatics Sciences (College of Agricultural and Life
Sciences) of the University of Florida.
Competing Interests
The authors declare there are no competing interests.
Author Contributions
Esteban N. Rodofili conceived and designed the experiments, performed the experiments,
analyzed the data, prepared figures and/or tables, authored or reviewed drafts of the
article, and approved the final draft.
Vincent Lecours performed the experiments, analyzed the data, authored or reviewed
drafts of the article, and approved the final draft.
Michelle LaRue performed the experiments, analyzed the data, authored or reviewed
drafts of the article, and approved the final draft.
Data Availability
The following information was supplied regarding data availability:
This work is a literature review and is not data-driven.
Supplemental Information
Supplemental information for this article can be found online at http://dx.doi.org/10.7717/
peerj.13540#supplemental-information.
REFERENCES
Abileah R. 2001. Use of high resolution space imagery to monitor the abundance,
distribution, and migration patterns of marine mammal populations. In: MTS/IEEE
Oceans 2001. An Ocean Odyssey. Conference Proceedings (IEEE Cat. No. 01CH37295).
Piscataway: IEEE, 1381–1387 DOI 10.1109/OCEANS.2001.968035.
Adame K, Pardo MA, Salvadeo C, Beier E, Elorriaga-Verplancken FR. 2017. Detectabil-
ity and categorization of California sea lions using an unmanned aerial vehicle.
Marine Mammal Science 33(3):913–925 DOI 10.1111/mms.12403.
Aniceto AS, Biuw M, Lindstrøm U, Solbø SA, Broms F, Carroll J. 2018. Monitoring
marine mammals using unmanned aerial vehicles: quantifying detection certainty.
Ecosphere 9(3):e02122 DOI 10.1002/ecs2.2122.
Arias-del Razo A, Heckel G, Schramm Y, Pardo MA. 2016. Terrestrial habitat prefer-
ences and segregation of four pinniped species on the islands off the western coast
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 16/22
of the Baja California Peninsula, Mexico. Marine Mammal Science 32(4):1416–1432
DOI 10.1111/mms.12339.
Bamford CCG, Kelly N, Dalla Rosa L, Cade DE, Fretwell PT, Trathan PN, Cubaynes
HC, Mesquita AFC, Gerrish L, Friedlaender AS, Jackson JA. 2020. A comparison
of baleen whale density estimates derived from overlapping satellite imagery and a
shipborne survey. Scientific Reports 10(1):12985 DOI 10.1038/s41598-020-69887-y.
Blaschke T, Hay GJ, Kelly M, Lang S, Hofmann P, Addink E, Queiroz Feitosa R, Van der
Meer F, Van der Werff H, Van Coillie F, Tiede D. 2014. Geographic object-based
image analysis—towards a new paradigm. ISPRS Journal of Photogrammetry and
Remote Sensing 87(100):180–191 DOI 10.1016/j.isprsjprs.2013.09.014.
Boehme L, Baker A, Fedak M, Årthun M, Nicholls K, Robinson P, Costa D, Biuw M,
Photopoulou T. 2016. Bimodal Winter Haul-Out Patterns of Adult Weddell Seals
(Leptonychotes weddellii) in the Southern Weddell Sea. PLOS ONE 11(5):e0155817
DOI 10.1371/journal.pone.0155817.
Born Riget FF, Dietz R, Andriashek D. 1999. Escape responses of hauled out ringed
seals (Phoca hispida) to aircraft disturbance. Polar Biology 21(3):171–178
DOI 10.1007/s003000050349.
Borowicz A, Le H, Humphries G, Nehls G, Höschle C, Kosarev V, Lynch HJ. 2019.
Aerial-trained deep learning networks for surveying cetaceans from satellite imagery.
PLOS ONE 14(10):e0212532 DOI 10.1371/journal.pone.0212532.
Burn DM, Cody MB. 2005. Use of satellite imagery to estimate walrus abundance
at Round Island, Alaska. In: Proceedings of the 16th Biennial Conference on the
Biology of Marine Mammals, San Diego, CA, USA.Available at https://www.
marinemammalscience.org/wp-content/uploads/2014/09/Abstracts-SMM-Biennial-
San-Diego-2005.pdf (accessed on 09 February 2022).
Charry B, Tissier E, Iacozza J, Marcoux M, Watt CA. 2021. Mapping Arctic cetaceans
from space: a case study for beluga and narwhal. PLOS ONE 16(8):e0254380
DOI 10.1371/journal.pone.0254380.
Clarke PJ, Cubaynes HC, Stockin KA, Olavarría C, De Vos A, Fretwell PT, Jackson JA.
2021. Cetacean strandings from space: challenges and opportunities of very high
resolution satellites for the remote monitoring of cetacean mass strandings. Frontiers
in Marine Science 8:650735 DOI 10.3389/fmars.2021.650735.
Corcoran E, Winsen M, Sudholz A, Hamilton G. 2021. Automated detection of wildlife
using drones: synthesis, opportunities and constraints. Methods in Ecology and
Evolution 12(6):1103–1114 DOI 10.1111/2041-210X.13581.
Corrêa AA, Quoos JH, Barreto AS, Groch KR, Eichler PPB. 2022. Use of satellite
imagery to identify southern right whales (Eubalaena australis) on a South-
west Atlantic Ocean breeding ground. Marine Mammal Science 38(1):87–101
DOI 10.1111/mms.12847.
Cubaynes HC. 2019. Whales from space: assessing the feasibility of using satellite imagery
to monitor whales. Doctoral thesis, University of Cambridge.
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 17/22
Cubaynes HC, Fretwell PT, Bamford C, Gerrish L, Jackson JA. 2019. Whales from
space: four mysticete species described using new VHR satellite imagery. Marine
Mammal Science 35(2):466–491 DOI 10.1111/mms.12544.
Davis GE, Baumgartner MF, Bonnell JM, Bell J, Berchok C, Bort Thornton J, Brault
S, Buchanan G, Charif RA, Cholewiak D, Clark CW, Corkeron P, Delarue J,
Dudzinski K, Hatch L, Hildebrand J, Hodge L, Klinck H, Kraus S, Martin B,
Mellinger DK, Moors-Murphy H, Nieukirk S, Nowacek DP, Parks S, Read AJ, Rice
AN, Risch D, Širović A, Soldevilla M, Stafford K, Stanistreet JE, Summers E, Todd
S, Warde A, Van Parijs SM. 2017. Long-term passive acoustic recordings track the
changing distribution of North Atlantic right whales (Eubalaena glacialis) from 2004
to 2014. Scientific Reports 7(1):13460 DOI 10.1038/s41598-017-13359-3.
Dawson S, Slooten E, DuFresne S, Wade P, Clement D. 2004. Small-boat surveys for
coastal dolphins: line-transect surveys for Hector’s dolphins (Cephalorhynchus
hectori). Fishery Bulletin 102(3):441–451.
Dujon AM, Ierodiaconou D, Geeson JJ, Arnould JPY, Allan BM, Katselidis KA,
Schofield G, Scales K, Bouchet P. 2021. Machine learning to detect marine animals
in UAV imagery: effect of morphology, spacing, behaviour and habitat. Remote
Sensing in Ecology and Conservation 7(3):341–354 DOI 10.1002/rse2.205.
Fiori L, Doshi A, Martinez E, Orams MB, Bollard-Breen B. 2017. The Use of un-
manned aerial systems in marine mammal research. Remote Sensing 9(6):543
DOI 10.3390/rs9060543.
Fischbach AS, Douglas DC. 2021. Evaluation of satellite imagery for monitor-
ing pacific walruses at a large Coastal Haulout. Remote Sensing 13(21):4266
DOI 10.3390/rs13214266.
Fretwell PT, Jackson JA, Ulloa Encina MJ, Häussermann V, Perez Alvarez MJ,
Olavarría C, Gutstein CS. 2019. Using remote sensing to detect whale strandings in
remote areas: the case of sei whales mass mortality in Chilean Patagonia. PLOS ONE
14(10):e0222498 DOI 10.1371/journal.pone.0222498.
Fretwell PT, Staniland IJ, Forcada J. 2014. Whales from space: counting southern right
whales by satellite. PLOS ONE 9(2):e88655 DOI 10.1371/journal.pone.0088655.
Goebel ME, Perryman WL, Hinke JT, Krause DJ, Hann NA, Gardner S, LeRoi DJ. 2015.
A small unmanned aerial system for estimating abundance and size of Antarctic
predators. Polar Biology 38(5):619–630 DOI 10.1007/s00300-014-1625-4.
Gon¸
calves BC, Spitzbart B, Lynch H. 2020. SealNet: a fully-automated pack-ice seal
detection pipeline for sub-meter satellite imagery. Remote Sensing of Environment
239:111617 DOI 10.1016/j.rse.2019.111617.
Gooday OJ, Key N, Goldstien S, Zawar-Reza P. 2018. An assessment of thermal-image
acquisition with an unmanned aerial vehicle (UAV) for direct counts of coastal
marine mammals ashore. Journal of Unmanned Vehicle Systems 6(2):100–108
DOI 10.1139/juvs-2016-0029.
Gray PC, Bierlich KC, Mantell SA, Friedlaender AS, Goldbogen JA, Johnston DW,
Ye H. 2019. Drones and convolutional neural networks facilitate automated and
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 18/22
accurate cetacean species identification and photogrammetry. Methods in Ecology
and Evolution 10(9):1490–1500 DOI 10.1111/2041-210X.13246.
Guirado E, Tabik S, Rivas ML, Alcaraz-Segura D, Herrera F. 2019. Whale counting
in satellite and aerial images with deep learning. Scientific Reports 9(1):14259
DOI 10.1038/s41598-019-50795-9.
Hashim NA, Jaaman S. 2011. Boat effects on the behaviour of Indo-Pacific Humpback
(Sousa chinensis) and Irrawaddy dolphins (Orcaella brevirostris) in Cowie Bay, Sabah,
Malaysia. Sains Malaysiana 40(12):1383–1392.
Hiby AR, Hammond PS. 1989. Survey techniques for estimating abundance of cetaceans.
Reports of the International Whaling Commission, Special Issue 11:47–80.
Hodgson A, Kelly N, Peel D. 2013. Unmanned aerial vehicles (UAVs) for surveying
marine fauna: a dugong case study. PLOS ONE 8(11):e79556
DOI 10.1371/journal.pone.0079556.
Hodgson A, Peel D, Kelly N. 2017. Unmanned aerial vehicles for surveying marine
fauna: assessing detection probability. Ecological Applications 27(4):1253–1267
DOI 10.1002/eap.1519.
Hollings T, Burgman M, Van Andel M, Gilbert M, Robinson T, Robinson A, McPher-
son J. 2018. How do you find the green sheep? A critical review of the use of re-
motely sensed imagery to detect and count animals. Methods in Ecology and Evolution
9(4):881–892 DOI 10.1111/2041-210X.12973.
Horton TW, Hauser N, Zerbini AN, Francis MP, Domeier ML, Andriolo A, Costa DP,
Robinson PW, Duffy CAJ, Nasby-Lucas N, Holdaway RN, Clapham PJ. 2017. Route
fidelity during marine megafauna migration. Frontiers in Marine Science 4:422
DOI 10.3389/fmars.2017.00422.
Horton TW, Zerbini AN, Andriolo A, Danilewicz D, Sucunza F. 2020. Multi-decadal
humpback whale migratory route fidelity despite oceanographic and geomagnetic
change. Frontiers in Marine Science 7:414 DOI 10.3389/fmars.2020.00414.
Höschle C, Cubaynes HC, Clarke PJ, Humphries G, Borowicz A. 2021. The potential of
satellite imagery for surveying whales. Sensors 21(3):963 DOI 10.3390/s21030963.
Johnston DW. 2019. Unoccupied aircraft systems in marine science and conservation.
Annual Review of Marine Science 11(1):439–463
DOI 10.1146/annurev-marine-010318-095323.
Koski WR, Abgrall P, Yazvenko SB. 2010. An inventory and evaluation of unmanned
aerial systems for offshore surveys of marine mammals. Journal of Cetacean Research
and Management 11(3):239–247.
LaRue MA, Ainley DG, Pennycook J, Stamatiou K, Salas L, Nur N, Stammerjohn S,
Barrington L, Horning N, Scales K. 2020. Engaging the crowd in remote sensing
to learn about habitat affinity of the Weddell seal in Antarctica. Remote Sensing in
Ecology and Conservation 6(1):70–78 DOI 10.1002/rse2.124.
LaRue MA, Rotella JJ, Garrott RA, Siniff DB, Ainley DG, Stauffer GE, Porter CC,
Morin PJ. 2011. Satellite imagery can be used to detect variation in abundance of
Weddell seals (Leptonychotes weddellii) in Erebus Bay, Antarctica. Polar Biology
34(11):1727–1737 DOI 10.1007/s00300-011-1023-0.
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 19/22
LaRue MA, Salas L, Nur N, Ainley DG, Stammerjohn S, Barrington L, Stamatious K,
Pennycook J, Doziers M, Saints J. Nakamura, H. 2019. Physical and ecological
factors explain the distribution of Ross Sea Weddell seals during the breeding season.
Marine Ecology Progress Series 612:193–208 DOI 10.3354/meps12877.
LaRue MA, Salas L, Nur N, Ainley D, Stammerjohn S, Pennycook J, Dozier M, Saints
J, Stamatiou K, Barrington L, Rotella J. 2021. Insights from the first global pop-
ulation estimate of Weddell seals in Antarctica. Science Advances 7(39):eabh3674
DOI 10.1126/sciadv.abh3674.
LaRue MA, Stapleton S. 2018. Estimating the abundance of polar bears on Wrangel
Island during late summer using high-resolution satellite imagery: a pilot study.
Polar Biology 41(12):2621–2626 DOI 10.1007/s00300-018-2384-4.
LaRue MA, Stapleton S, Anderson M. 2017. Feasibility of using high-resolution
satellite imagery to assess vertebrate wildlife populations. Conservation Biology
31(1):213–220 DOI 10.1111/cobi.12809.
LaRue MA, Stapleton S, Porter C, Atkinson S, Atwood T, Dyck M, Lecomte N.
2015. Testing methods for using high-resolution satellite imagery to monitor
polar bear abundance and distribution. Wildlife Society Bulletin 39(4):772–779
DOI 10.1002/wsb.596.
Linchant J, Lisein J, Semeki J, Lejeune P, Vermeulen C. 2015. Are unmanned aircraft
systems (UASs) the future of wildlife monitoring? A review of accomplishments and
challenges. Mammal Review 45(4):239–252 DOI 10.1111/mam.12046.
Lu D, Mausel P, Brondízio E, Moran E. 2004. Change detection techniques. International
Journal of Remote Sensing 25(12):2365–2401 DOI 10.1080/0143116031000139863.
Luksenburg J, Parsons ECM. 2009. Effects of aircraft on cetaceans: implications for
aerial whalewatching. In: Proceedings of the 61st meeting of the international whaling
commission. Madeira. Available at https://www.researchgate.net/publication/233747059
(accessed on 09 February 2022).
Maire F, Mejias L, Hodgson A, Duclos G. 2013. Detection of dugongs from unmanned
aerial vehicles. In: Proceedings of the IEEE/RSJ international conference on intelligent
robots and systems. Tokyo, 2750–2756 DOI 10.1109/IROS.2013.6696745.
Mejias Alvarez L, Duclos G, Hodgson A, Maire F. 2013. Automated marine mammal
detection from aerial imagery. In: Wernli Sr R, ed. Proceedings of the OCEANS ’13
conference. Institute of Electrical and Electronics Engineers Inc, 1–5. Available at
https://eprints.qut.edu.au/61589/ (accessed on 09 February 2022).
Merrick RL, Silber GK, De Master DP. 2018. Endangered Species and Populations. In:
Würsig B, Thewissen JGM, Kovacs KM, eds. Encyclopedia of Marine Mammals. Saint
Louis: Elsevier Science & Technology, 313–318.
Moore JE, Forney KA, Weller DW. 2018. Surveys. In: Würsig B, Thewissen JGM,
Kovacs KM, eds. Encyclopedia of Marine Mammals. Saint Louis: Elsevier Science &
Technology, 960–963.
Platonov NG, Mordvintsev IN, Rozhnov VV. 2013. The possibility of using high
resolution satellite images for detection of marine mammals. Biology Bulletin of the
Russian Academy of Sciences 40(2):197–205 DOI 10.1134/S1062359013020106.
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 20/22
Pomeroy P, O’Connor L, Davies P. 2015. Assessing use of and reaction to unmanned
aerial systems in gray and harbor seals during breeding and molt in the UK. Journal
of Unmanned Vehicle Systems 3(3):102–113 DOI 10.1139/juvs-2015-0013.
Raoult V, Colefax AP, Allan BM, Cagnazzi D, Castelblanco-Martínez N, Ierodiaconou
D, Johnston DW, Landeo-Yauri S, Lyons M, Pirotta V, Schofield G, Butcher PA.
2020. Operational protocols for the use of drones in marine animal research. Drones
4(4):64 DOI 10.3390/drones4040064.
Sasse DB. 2003. Job-related mortality of wildlife workers in the United States, 1937-2000.
Wildlife Society Bulletin 31(4):1015–1020.
Schofield G, Esteban N, Katselidis KA, Hays GC. 2019. Drones for research on sea
turtles and other marine vertebrates—a review. Biological Conservation 238:108214
DOI 10.1016/j.biocon.2019.108214.
Seymour A, Dale J, Hammill M, Halpin PN, Johnston DW. 2017. Automated detection
and enumeration of marine wildlife using unmanned aircraft systems (UAS) and
thermal imagery. Scientific Reports 7(1):45127 DOI 10.1038/srep45127.
Singh A. 1989. Review Article Digital change detection techniques using remotely-sensed
data. International Journal of Remote Sensing 10(6):989–1003
DOI 10.1080/01431168908903939.
Smith CE, Sykora-Bodie ST, Bloodworth B, Pack SM, Spradlin TR, LeBoeuf NR.
2016. Assessment of known impacts of unmanned aerial systems (UAS) on marine
mammals: data gaps and recommendations for researchers in the United States.
Journal of Unmanned Vehicle Systems 4(1):31–44 DOI 10.1139/juvs-2015-0017.
Stanistreet JE, Nowacek D, Bell J, Cholewiak D, Hildebrand J, Hodge L, Van Parijs S,
Read A. 2018. Spatial and seasonal patterns in acoustic detections of sperm whales
Physeter macrocephalus along the continental slope in the western North Atlantic
Ocean. Endangered Species Research 35:1–13 DOI 10.3354/esr00867.
Subhan B, Arafat D, Santoso P, Pahlevi K, Prabowo B, Taufik M, Kusumo BS, Awak
K, Khaerudi D, Ohoiulun H, Nasetion FI, Madduppa H. 2019. Development of
observing dolphin population method using Small Vertical Take-off and Landing
(VTOL) Unmanned Aerial System (AUV). IOP Conference Series. Earth and
Environmental Science 278(1):012074 DOI 10.1088/1755-1315/278/1/012074.
Thums M, Jenner C, Newland C, Ferreira L, Waples K, Meekan M. 2018. Chapter
2: Detecting humpback whales from high resolution satellite imagery (2018). In
Thums M., Jenner C., Waples K., Salgado Kent C. & Meekan M. Humpback whale
use of the Kimberley; understanding and monitoring spatial distribution. Report
of Project 1.2.1 prepared for the Kimberley Marine Research Program, Western
Australian Marine Science Institution, Perth, Western Australia. 27–36. Available
at https://wamsi.org.au/wp-content/uploads/bsk-pdf-manager/2019/07/Final-Report-
WAMSI-KMRP-Whales-Humpback-Whale-Use-of-the-Kimberley-Thums-et-al.pdf
(accessed on 08 February 2022).
Turner W, Rondinini C, Pettorelli N, Mora B, Leidner A, Szantoi Z, Buchanan
G, Dech S, Dwyer J, Herold M, Koh L., Leimgruber P, Taubenboeck H, Weg-
mann M, Wikelski M, Woodcock C. 2015. Free and open-access satellite data
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 21/22
are key to biodiversity conservation. Biological Conservation 182:173–176
DOI 10.1016/j.biocon.2014.11.048.
UF George A. Smathers Libraries. 2022. Databases. Available at https://uflib.ufl.edu/find/
databases/ (accessed on 25 April 2022).
United States Geological Survey. 2020. What are the band designations for the Landsat
satellites? Available at www.usgs.gov/faqs/what-are-band-designations-landsat-
satellites?qt-news_science_products=0# qt-news_science_products (accessed on 08
February 2022).
Wang D, Shao Q, Yue H. 2019. Surveying wild animals from satellites, manned aircraft
and Unmanned Aerial Systems (UASs): a review. Remote Sensing 11(11):1308
DOI 10.3390/rs11111308.
Würsig B, Lynn SK, Jefferson TA, Mullin KD. 1998. Behaviour of cetaceans in the
Northern Gulf of Mexico relative to survey ships and aircraft. Aquatic Mammals
24(1):41–50.
Ye S, Pontius RG, Rakshit R. 2018. A review of accuracy assessment for object-based
image analysis: from per-pixel to per-polygon approaches. ISPRS Journal of Pho-
togrammetry and Remote Sensing 141:137–147 DOI 10.1016/j.isprsjprs.2018.04.002.
Zinglersen KB, Garde E, Langley K, Mätzler E. 2020. RemoteID. Identification of
Atlantic Walrus at haul out sites in Greenland using high-resolution satellite images.
Technical report (111) Greenland Institute of Natural Resources, Greenland. Avail-
able at https://natur.gl/wp-content/uploads/2020/03/111-RemoteID.-Identification-of-
Atlantic-Walrus-at-haul-out-sites.pdf (accessed on 09 February 2022).
Rodofili et al. (2022), PeerJ, DOI 10.7717/peerj.13540 22/22
... The automation of marine mammal detection in remote sensing imagery (including UAS imagery) constitutes a complementary tool to manual counts worth exploring; it could help save researcher time, effort (e.g., reducing human fatigue from manual analyses), and cost (Rodofili et al., 2022). This is particularly relevant when animals are found over large areas or in large aggregations, like those formed by Florida manatees in thermal refugia during the colder months (Reynolds et al., 2018). ...
... This is particularly relevant when animals are found over large areas or in large aggregations, like those formed by Florida manatees in thermal refugia during the colder months (Reynolds et al., 2018). However, we found relatively few works on automated detection of sirenians through UAS imagery analysis-mostly in conference papers on dugongs (see Table 1 in Rodofili et al., 2022). For example, Mejias applied an automated workflow to detect dugongs in UAS imagery, obtaining recalls (Equation 1) of 48.57 ...
... Image analysis was based on OBIA principles, which have not been explored extensively for marine mammal detection compared to more traditional pixel-based methods (Rodofili et al., 2022). These two approaches are fundamentally different in their units of analysis (Blaschke et al., 2014). ...
Article
Full-text available
Florida manatees (Trichechus manatus latirostris) require frequent and extensive surveys to inform conservation efforts. Crewed aircraft surveys can be costly, dangerous, and logistically complex. Unoccupied aerial systems (UASs) can assist with these issues. While manual review of UAS imagery can be time- and labor-intensive, automated detection of manatees in aerial survey footage can help. We present an object-based image analysis workflow for the automated detection and count of Florida manatees in Google Earth Engine, a free platform for research that allows for scripts and imagery sharing. Training and testing datasets were built from randomly extracted image frames from two stationary, unoccupied aerial system videos over thermal refugia. The workflow captured most manatees (93.98 to 95.62% recall; 4.38 to 6.03% false negative rate), but also counted many objects as manatees incorrectly (4.24 to 14.77% precision; 998.40 to 3,885.54% false positive over the detectable rate). Sun glint, mud plumes, and water close to shore were common causes of false positives. While the automated count was too high, the workflow lays markers over each detection, allowing for quick manual review for more accurate (semi-automated) counts. This study is an early step in automated detection tools for Florida manatees in a cloud-based platform. Future efforts could explore other platforms or may improve this workflow by including new classes for confounding objects.
... Several metrics, including the false negative rate (FNR), false positive rate (FPR) and automated count deviation (ACD) (following [31]), were calculated based on these counts. These metrics were used to assess the accuracy and reliability of each algorithm for both the training and test tiles and provide a more direct assessment of the accuracy and reliability of the semiautomated detection algorithm [31]. ...
... Several metrics, including the false negative rate (FNR), false positive rate (FPR) and automated count deviation (ACD) (following [31]), were calculated based on these counts. These metrics were used to assess the accuracy and reliability of each algorithm for both the training and test tiles and provide a more direct assessment of the accuracy and reliability of the semiautomated detection algorithm [31]. The false negative rate (or missed rate) provides a direct measure of the undercount for each algorithm. ...
Article
Full-text available
Very high-resolution (VHR) satellite imagery has proven to be useful for detection of large to medium cetaceans, such as odontocetes and offers some significant advantages over traditional detection methods. However, the significant time investment needed to manually read satellite imagery is currently a limiting factor to use this method across large open ocean regions. The objective of this study is to develop a semi-automated detection method using object-based image analysis to identify beluga whales (Delphinapterus leucas) in open water (summer) ocean conditions in the Arctic using panchromatic WorldView-3 satellite imagery and compare the detection time between human read and algorithm detected imagery. The false negative rate, false positive rate, and automated count deviation were used to assess the accuracy and reliability of various algorithms for reading training and test imagery. The best algorithm, which used spectral mean and texture variance attributes, detected no false positives and the false negative rate was low (<4%). This algorithm was able to accurately and reliably identify all the whales detected by experienced readers in the ice-free panchromatic image. The autodetection algorithm does have difficulty separately identifying whales that are perpendicular to one another, whales below the surface, and may use multiple segments to define a whale. As a result, for determining counts of whales, a reader should manually review the automated results. However, object-based image analysis offers a viable solution for processing large amounts of satellite imagery for detecting medium-sized beluga whales while eliminating all areas of the imagery which are whale-free. This algorithm could be adapted for detecting other cetaceans in ice-free water.
... The count deviation (3) (Rodofili et al., 2022), on the whale class, was calculated to provide a measure of the cumulative mistakes made by the model as a fraction of the total number of samples. ...
... Most ambitious would be to generate results which are informative about absolute whale density (e.g., Bamford et al., 2020), for example, to measure local abundance. In the latter cases, it is particularly important to minimize bias generated by poor precision or recall metrics (Rodofili et al., 2022). The mAP@0.5 is the metric used to judge overall model performance and is maximized by the model in this study, but consideration should be given to the importance of each metric, particularly as the confidence at which the precision and recall are calculated is different from that of the mAP. ...
Article
Full-text available
The combination of very high resolution (VHR) satellite remote sensing imagery and deep learning via convolutional neural networks provides opportunities to improve global whale population surveys through increasing efficiency and spatial coverage. Many whale species are recovering from commercial whaling and face multiple anthropogenic threats. Regular, accurate population surveys are therefore of high importance for conservation efforts. In this study, a state‐of‐the‐art object detection model (YOLOv5) was trained to detect gray whales ( Eschrichtius robustus ) in VHR satellite images, using training data derived from satellite images spanning different sea states in a key breeding habitat, as well as aerial imagery collected by unoccupied aircraft systems. Varying combinations of aerial and satellite imagery were incorporated into the training set. Mean average precision, whale precision, and recall ranged from 0.823 to 0.922, 0.800 to 0.939, and 0.843 to 0.889, respectively, across eight experiments. The results imply that including aerial imagery in the training data did not substantially impact model performance, and therefore, expansion of representative satellite datasets should be prioritized. The accuracy of the results on real‐world data, along with short training times, indicates the potential of using this method to automate whale detection for population surveys.
... Resulting data could be used to validate results of model predictions like the ones presented in this study, and also provide baselines for future population forecasts. Manned helicopter surveys combined with manual image analysis (as done in Svalbard previously; Krafft, Kovacs, Andersen, et al., 2006) are costly, logistically challenging, and labor-intensive, but new alternatives combining unmanned aircraft systems, remote sensing, and machine learning are showing great promise (Rodofili et al., 2022;Seymour et al., 2017). Thermal imagery in particular has been shown to be suitable for monitoring abundance of ice-affiliated pinnipeds (Seymour et al., 2017;Young et al., 2019) as long as observation error is properly accounted for (Conn et al., 2014); simulation studies can help with designing optimized surveys (Conn et al., 2016). ...
Article
Full-text available
Throughout the Arctic, ice‐affiliated marine mammals constitute local subsistence resources but detrimental effects of declines in their sea ice habitats create a need for harvest sustainability assessments in light of climate change. At the same time, empirical data required for thorough population analysis of these species are often sparse at best, as illustrated by the focal species in this study, ringed seals in Svalbard: the last population survey took place two decades ago (2002–2003), demographic data are limited to age, sex, and reproductive status of a small subset of shot individuals, and harvest reporting is patchy and incomplete. Data sparsity is one of the main reasons why potential biological removal (PBR) became a commonly used tool for assessing sustainability of marine mammal harvests. Herein, we calculated PBR for Svalbard ringed seals using both recommended default parameters and population‐specific parameters obtained from an integrated population model (IPM). PBR estimates were highly uncertain, suggesting the number of sustainably harvestable individuals could lie anywhere between 0 and 91, with a substantial chance of any harvest being unsustainable under current environmental conditions and trends. Subsequent population viability analyses (PVAs) further confirmed that the current harvest was likely unsustainable, even in a scenario in which sea ice conditions would not deteriorate (and therefore lower pup survival) further. However, uncertainty in population projections was high, and forecasts thus not ideal for formulating management advice. Better forecasts will require more frequent population surveys and obtaining more knowledge regarding the links between vital rates and environmental conditions, both of which may be facilitated by the adoption of novel technology (e.g., drone monitoring, genetic studies). The modeling framework created in this study can be readily updated with new data as they become available, and can serve as a tool for adaptive management of this and other marine mammal populations.
... Localization and identification of objects of interest in remote sensing images (RSIs) are essential for object identification, resource management, decision-making, and disaster relief response. It plays a catalytic role in military reconnaissance (Xie et al., 2024), ecological protection (Rodofili, Lecours & LaRue, 2022), unmanned vehicles (Lyu et al., 2022), urban planning (Shen et al., 2023) to name a few. Although many researchers have proposed many algorithms for remote sensing object detection (RSOD), this task still needs to overcome many difficulties, mainly due to the complex backgrounds, dense target quantities, large-scale variations, and small-scale objects. ...
Article
Full-text available
Accurate localization of objects of interest in remote sensing images (RSIs) is of great significance for object identification, resource management, decision-making and disaster relief response. However, many difficulties, like complex backgrounds, dense target quantities, large-scale variations, and small-scale objects, which make the detection accuracy unsatisfactory. To improve the detection accuracy, we propose an Adaptive Adjacent Context Negotiation Network (A ² CN-Net). Firstly, the composite fast Fourier convolution (CFFC) module is given to reduce the information loss of small objects, which is inserted into the backbone network to obtain spectral global context information. Then, the Global Context Information Enhancement (GCIE) module is given to capture and aggregate global spatial features, which is beneficial for locating objects of different scales. Furthermore, to alleviate the aliasing effect caused by the fusion of adjacent feature layers, a novel Adaptive Adjacent Context Negotiation network (A ² CN) is given to adaptive integration of multi-level features, which consists of local and adjacent branches, with the local branch adaptively highlighting feature information and the adjacent branch introducing global information at the adjacent level to enhance feature representation. In the meantime, considering the variability in the focus of feature layers in different dimensions, learnable weights are applied to the local and adjacent branches for adaptive feature fusion. Finally, extensive experiments are performed in several available public datasets, including DIOR and DOTA-v1.0. Experimental studies show that A ² CN-Net can significantly boost detection performance, with mAP increasing to 74.2% and 79.2%, respectively.
... Occupied aircraft have been used to survey and estimate pinniped populations since the era of industrial sealing (Bartlett 1929), exploiting aerial perspectives to scout large regions of land or ice habitat at a time. Today, high-resolution satellites provide even greater spatial coverage of pinniped habitats (LaRue et al. 2011, Rodofili et al. 2022, with advantages that include automated and relatively passive data collection, once sensors are placed in orbit, and regular coverage that depends on the satellite's orbit and revisit period, though this is reduced by coincident cloud cover. Imagery from occupied aircraft regularly achieves GSDs and quality necessary to distinguish seals in their ice or land habitats and, under select circumstances, very high-resolution satellite imagery can enable the same (LaRue et al. 2017). ...
Article
Full-text available
Pinniped species undergo uniquely amphibious life histories that make them valuable subjects for many domains of research. Pinniped research has often progressed hand‐in‐hand with technological frontiers of wildlife biology, and drones represent a leap forward for methods of aerial remote sensing, enabling data collection, and integration at new scales of biological importance. Drone methods and data types provide four key opportunities for wildlife surveillance that are already advancing pinniped research and management: 1) repeat and on‐demand surveillance, 2) high‐resolution coverage at large extents, 3) morphometric photogrammetry, and 4) computer vision and deep learning applications. Drone methods for pinniped research represent early stages of technological adoption and can reshape the field as they scale towards the full potential of their techniques.
... Future work would advance from exploring ways to include such information in the analysis. Additionally, next steps should explore ways to efficiently increase the sample size of data, possibly, through the automation of vessel detection which seems viable given the high spectral contrast between the fiberglass hull of the majority of vessels and the sea surface; marine mammal detection has been successfully implemented before despite the spectral similarities between the animals and the water surface (Rodofili et al., 2022), making them theoretically more difficult to automatically detect than vessels. Automation of vessel detection and use of a wider set of satellites as observation platforms, will enable increased amounts of satellite images to be analyzed potentially confirming the significance of the presented results and revealing further details regarding the spatio-temporal distribution of vessels in the study area. ...
Article
Characterizing specific pressures on natural resources is a key component of developing sustainable management strategies. Here, we explored the spatiotemporal distribution of marine vessels in a potentially high-conflict area between humpback whales and tourism activities, namely Bahía de Banderas, Mexico, from 2017 to 2023. Using high-resolution satellite imagery, we identified high-use areas and their relationship to the presence of urban areas, vessel departure points, and the coastline before, during, and after restrictions put in place during the peak of the COVID-19 pandemic. High-resolution satellite imagery (3m/px - 5m/px) retrieved through the Planet Application Program was used to perform manual detection and annotation of vessels, and spatial patterns of marine navigation were characterized through point pattern analyses. The spatial distribution of all marine vessel observations in the study area was not random; four high-use coastal areas closer to urban areas were identified, three of which overlap the current marine protected area established to protect whale nursing grounds. Furthermore, as the COVID-19 pandemic began, coastal navigation decreased due to lockdown before picking up again at a higher level than the pre-COVID baseline due to social distancing restrictions. While these changes were not statistically significant, they potentially indicate a behavioral change caused by COVID-19 social distancing recommendations during 2021; the amplified demand for marine vessels appears to have increased the absolute number of vessel chartering and decreased the number of passengers onboard commercial vessels. Our preliminary work provides a reproducible remote sensing methodology that can support future mitigation strategies.
... Satellite imagery offers a safe, effective method for estimating cetacean estuary abundance; however, ground-truthing with traditional aerial surveys would improve confidence in the estimate (Bamford et al., 2020). Crowd-counters offer one possible solution to handling the immense imagery reading task, but the future of abundance estimation with satellite imagery likely relies on automated detection algorithms that can be used to eliminate large areas with no features of interest and identify whales (Borowicz, 2019;Rodofili et al., 2022;Khan et al., 2023), while also eliminating the subjectivity of readers. Object or pixel based machine learning algorithms could greatly reduce the time required for analysis after imagery acquisition. ...
Article
Full-text available
Introduction The Eastern High Arctic–Baffin Bay (EHA-BB) beluga whale ( Delphinapterus leucas ) population spends summer in estuaries around Somerset Island, Nunavut, Canada. A single abundance estimate from 1996 suggests an abundance >21,000 beluga whales; however, more information on abundance and distribution is needed to ensure effective management of this population, especially in estuaries where previous surveys provided minimal coverage. To assess the feasibility of using Very High Resolution (VHR) satellite imagery to obtain estuary abundance estimates for this beluga population, we evaluated a citizen science crowd counting initiative that was designed to monitor remote beluga whale populations and their estuary use. Methods In July and August 2020 the WorldView 2 and 3, and GeoEye 1 satellites were tasked to collect VHR imagery (30–41 cm) of estuaries previously known to be used by Eastern High Arctic–Baffin Bay beluga whales. The objectives were to obtain an estuary abundance estimate for this population from satellite imagery, and to evaluate the effectiveness of having imagery annotated using a crowd-source platform. Almost 3,800 km ² of ocean imagery was analyzed using Maxar’s Geospatial Human Imagery Verification Effort (GeoHIVE) Crowdsourcing platform. Expert readers then manually compared counts to those performed by crowd-counters to determine variance in observer counts. Results and Discussion The estuary abundance estimate from 11 core estuaries was 12,128 (CV 36.76%, 95% confidence interval 6,036–24,368) beluga whales. This represents an estuary abundance estimate only, as the greater Peel Sound and Prince Regent Inlet areas were not photographed. The estuaries with the largest abundance of beluga whales were Creswell Bay, Maxwell Bay, and Prince Whales Island, with over 2,000 crowd-counted whales in each estuary. Although VHR imagery has potential to assist with surveying and monitoring marine mammals, for larger estuaries it was not always possible to photograph the entire area in a single day, and cloud cover was an issue for sections of most images. This work will assist with planning large-scale aerial surveys for monitoring beluga whale populations, identifying high-use areas and important beluga habitat, and highlights the utility of using VHR imagery to enhance our understanding of estuary abundance and distribution of Arctic whales.
Article
Satellite images could aid in studying baleen whale migration routes given their extensive spatial coverage. We developed two object-based image analysis (OBIA) workflows to automatically detect migrating whales and their orientation. We used three sections of a Worldview-3 satellite image off California with migrating gray whales (Eschrichtius robustus), alternating training and testing under a three-fold approach in eCognition, with three classes: whale, water, and confounding objects (e.g., sea foam). After a workflow with 202 classification features, we applied feature selection through a minimum redundancy maximum relevance algorithm and retained 15 features. This selection was used in a second workflow which yielded fewer false negatives (FNs) and especially, fewer false positives (FPs)(22.6% vs. 11.1% FNs rate, 7599.6% vs. 1352.6% FP over detectable rate). While FPs were still considerable, image grid cells to review were reduced to less than 3% of cells for full manual analysis. Using Moore's test for paired circular data, we failed to find significant differences between manual and both automated or semiautomated orientation measurements (corrected for whales measured from head to tail). Most whales were oriented in the southeastern quadrant. The results are promising for satellite image studies of migration routes.
Preprint
Pinniped species undergo uniquely amphibious life histories that make them valuable subjects for many domains of research. Pinniped research has often progressed hand-in-hand with technological frontiers of wildlife biology, and drones represent a leap forward for methods of aerial remote sensing, heralding data collection and integration at new scales of biological importance. Drone methods and data types provide four key opportunities for wildlife surveillance that are already advancing pinniped research and management: (1) repeat and on demand surveillance, (2) high-resolution coverage at large extents, (3) morphometric photogrammetry, and (4) computer vision and deep learning applications. Drone methods for pinniped research represent early stages of technological adoption and can reshape the field as they scale towards the full potential of their techniques.
Article
Full-text available
The study of cetacean strandings was globally recognised as a priority topic at the 2019 World Marine Mammal Conference, in recognition of its importance for understanding the threats to cetacean communities and, more broadly, the threats to ecosystem and human health. Rising multifaceted anthropogenic and environmental threats across the globe, as well as whale population recovery from exploitation in some areas, are likely to coincide with an increase in reported strandings. However, the current methods to monitor strandings are inherently biased towards populated coastlines, highlighting the need for additional surveying tools in remote regions. Very High Resolution (VHR) satellite imagery offers the prospect of upscaling monitoring of mass strandings in minimally populated/unpopulated and inaccessible areas, over broad spatial and temporal scales, supporting and informing intervention on the ground, and can be used to retrospectively analyse historical stranding events. Here we (1) compile global strandings information to identify the current data gaps; (2) discuss the opportunities and challenges of using VHR satellite imagery to monitor strandings using the case study of the largest known baleen whale mass stranding event (3) consider where satellites hold the greatest potential for monitoring strandings remotely and; (4) outline a roadmap for satellite monitoring. To utilise this platform to monitor mass strandings over global scales, considerable technical, practical and environmental challenges need to be addressed and there needs to be inclusivity in opportunity from the onset, through knowledge sharing and equality of access to imagery.
Article
Full-text available
Pacific walruses (Odobenus rosmarus divergens) are using coastal haulouts in the Chukchi Sea more often and in larger numbers to rest between foraging bouts in late summer and autumn in recent years, because climate warming has reduced availability of sea ice that historically had provided resting platforms near their preferred benthic feeding grounds. With greater numbers of walruses hauling out in large aggregations, new opportunities are presented for monitoring the population. Here we evaluate different types of satellite imagery for detecting and delineating the peripheries of walrus aggregations at a commonly used haulout near Point Lay, Alaska, in 2018–2020. We evaluated optical and radar imagery ranging in pixel resolutions from 40 m to ~1 m: specifically, optical imagery from Landsat, Sentinel-2, Planet Labs, and DigitalGlobe, and synthetic aperture radar (SAR) imagery from Sentinel-1 and TerraSAR-X. Three observers independently examined satellite images to detect walrus aggregations and digitized their peripheries using visual interpretation. We compared interpretations between observers and to high-resolution (~2 cm) ortho-corrected imagery collected by a small unoccupied aerial system (UAS). Roughly two-thirds of the time, clouds precluded clear optical views of the study area from satellite. SAR was unaffected by clouds (and darkness) and provided unambiguous signatures of walrus aggregations at the Point Lay haulout. Among imagery types with 4–10 m resolution, observers unanimously agreed on all detections of walruses, and attained an average 65% overlap (sd 12.0, n 100) in their delineations of aggregation boundaries. For imagery with ~1 m resolution, overlap agreement was higher (mean 85%, sd 3.0, n 11). We found that optical satellite sensors with moderate resolution and high revisitation rates, such as PlanetScope and Sentinel-2, demonstrated robust and repeatable qualities for monitoring walrus haulouts, but temporal gaps between observations due to clouds were common. SAR imagery also demonstrated robust capabilities for monitoring the Point Lay haulout, but more research is needed to evaluate SAR at haulouts with more complex local terrain and beach substrates.
Article
Full-text available
The Weddell seal is one of the best-studied marine mammals in the world, owing to a multidecadal demographic effort in the southernmost part of its range. Despite their occurrence around the Antarctic coastline, we know little about larger scale patterns in distribution, population size, or structure. We combined high-resolution satellite imagery from 2011, crowd-sourcing, and habitat modeling to report the first global population estimate for the species and environmental factors that influence its distribution. We estimated ~202,000 (95% confidence interval: 85,345 to 523,140) sub-adult and adult female seals, with proximate ocean depth and fast-ice variables as factors explaining spatial prevalence. Distances to penguin colonies were associated with seal presence, but only emperor penguin population size had a strong negative relationship. The small, estimated population size relative to previous estimates and the seals’ nexus with trophic competitors indicates that a community ecology approach is required in efforts to monitor the Southern Ocean ecosystem.
Article
Full-text available
Emergence of new technologies in remote sensing give scientists a new way to detect and monitor wildlife populations. In this study we assess the ability to detect and classify two emblematic Arctic cetaceans, the narwhal (Monodon monoceros) and beluga whale (Delphinapterus leucas), using very high-resolution (VHR) satellite imagery. We analyzed 12 VHR images acquired in August 2017 and 2019, collected by the WorldView-3 satellite, which has a maximum resolution of 0.31 m per pixel. The images covered Clearwater Fiord (138.8 km²), an area on eastern Baffin Island, Canada where belugas spend a large part of the summer, and Tremblay Sound (127.0 km²), a narrow water body located on the north shore of Baffin Island that is used by narwhals during the open water season. A total of 292 beluga whales and 109 narwhals were detected in the images. This study contributes to our understanding of Arctic cetacean distribution and highlights the capabilities of using satellite imagery to detect marine mammals.
Article
Full-text available
Abstract Machine learning algorithms are being increasingly used to process large volumes of wildlife imagery data from unmanned aerial vehicles (UAVs); however, suitable algorithms to monitor multiple species are required to enhance efficiency. Here, we developed a machine learning algorithm using a low‐cost computer. We trained a convolutional neural network and tested its performance in: (1) distinguishing focal organisms of three marine taxa (Australian fur seals, loggerhead sea turtles and Australasian gannets; body size ranges: 0.8–2.5 m, 0.6–1.0 m, and 0.8–0.9 m, respectively); and (2) simultaneously delineating the fine‐scale movement trajectories of multiple sea turtles at a fish cleaning station. For all species, the algorithm performed best at detecting individuals of similar body length, displaying consistent behaviour or occupying uniform habitat (proportion of individuals detected, or recall of 0.94, 0.79 and 0.75 for gannets, seals and turtles, respectively). For gannets, performance was impacted by spacing (huddling pairs with offspring) and behaviour (resting vs. flying shapes, overall precision: 0.74). For seals, accuracy was impacted by morphology (sexual dimorphism and pups), spacing (huddling and creches) and habitat complexity (seal sized boulders) (overall precision: 0.27). For sea turtles, performance was impacted by habitat complexity, position in water column, spacing, behaviour (interacting individuals) and turbidity (overall precision: 0.24); body size variation had no impact. For sea turtle trajectories, locations were estimated with a relative positioning error of
Article
Full-text available
The emergence of very high-resolution (VHR) satellite imagery (less than 1 m spatial resolution) is creating new opportunities within the fields of ecology and conservation biology. The advancement of sub-meter resolution imagery has provided greater confidence in the detection and identification of features on the ground, broadening the realm of possible research questions. To date, VHR imagery studies have largely focused on terrestrial environments; however, there has been incremental progress in the last two decades for using this technology to detect cetaceans. With advances in computational power and sensor resolution, the feasibility of broad-scale VHR ocean surveys using VHR satellite imagery with automated detection and classification processes has increased. Initial attempts at automated surveys are showing promising results, but further development is necessary to ensure reliability. Here we discuss the future directions in which VHR satellite imagery might be used to address urgent questions in whale conservation. We highlight the current challenges to automated detection and to extending the use of this technology to all oceans and various whale species. To achieve basin-scale marine surveys, currently not feasible with any traditional surveying methods (including boat-based and aerial surveys), future research requires a collaborative effort between biology, computation science, and engineering to overcome the present challenges to this platform’s use.
Thesis
Full-text available
By the mid-twentieth century, the majority of great whale species were threatened with extinction, following centuries of commercial whaling. Since the implementation of a moratorium on commercial whaling in 1985 by the International Whaling Commission, the recovery of whale population is being regularly assessed. Various methods are used to survey whale populations, though most are spatially limited and prevent remote areas from being studied. Satellites orbiting Earth can access most regions of the planet, offering a potential solution to surveying remote locations. With recent improvements in the spatial resolution of satellite imagery, it is now possible to detect wildlife from space, including whales. In this thesis, I aimed to further investigate the feasibility of very high resolution (VHR) satellite imagery as a tool to reliably monitor whales. The first objective was to describe, both visually and spectrally, how four morphologically distinct species appear in VHR satellite imagery. The second objective was to explore different ways to automatically detect whales in such imagery, as the current alternative is manual detection, which is time-consuming and impractical when monitoring large areas. With the third objective, I attempted to give some insights on how to estimate the maximum depth at which a whale can be detected in VHR satellite imagery, as this will be crucial to estimate whale abundance from space. This thesis shows that the four species targeted could be detected with varying degrees of accuracy, some contrasting better with their surroundings. Compared to manual detection, the automated systems trialled here took longer, were not as accurate, and were not transferable to other images, suggesting to focus future automation research on machine learning and the creation of a well-labelled database required to train and validate. The maximum depth of detection could be assessed only approximately using nautical charts. Other methods such as the installation of panels at various depths should be trialled, although it requires prior knowledge of the spectral reflectance of whales above the surface, which I tested on post-mortem samples of whale integument and proved unreliable. Such reflectance should be measured on free-swimming whale using unmanned aerial vehicles or small aircraft. Overall, this thesis shows that currently VHR satellite imagery can be a useful tool to assess the presence or absence of whales, encouraging further developments to make VHR satellite imagery a reliable method to monitor whale numbers.
Article
Full-text available
The use of drones to study marine animals shows promise for the examination of numerous aspects of their ecology, behaviour, health and movement patterns. However, the responses of some marine phyla to the presence of drones varies broadly, as do the general operational protocols used to study them. Inconsistent methodological approaches could lead to difficulties comparing studies and can call into question the repeatability of research. This review draws on current literature and researchers with a wealth of practical experience to outline the idiosyncrasies of studying various marine taxa with drones. We also outline current best practice for drone operation in marine environments based on the literature and our practical experience in the field. The protocols outlined herein will be of use to researchers interested in incorporating drones as a tool into their research on marine animals and will help form consistent approaches for drone-based studies in the future.
Article
Satellite imagery has been used to improve scientific research worldwide. In this study, the southern right whale (Eubalaena australis) was chosen to test the use of medium, high, and very high resolution (VHR) satellite images, on the Brazilian breeding ground. These images were used to identify the whales and were compared to aerial survey data collected in the same area. The VHR satellite images from the Pleiades‐1A satellite, available on Google Earth, displayed the best results when compared to those from Sentinel 2, Landsat 8, Rapid Eye, and Planet Scope. No significant differences were observed (Mann‐Whitney U test) between the possible whales recorded in the satellite images and of real whales recorded in situ by aerial surveys, while considering either the number of groups (p = .841, n = 5) or the total number of animals (p = .222, n = 5). Further, when using VHR images, the geographical positions of the whales recorded in situ had a positive correlation with the positions generated from the satellite images (Mantel test: r = 0.52, p = .001, n = 13). This technique may represent an important tool for detecting right whales, especially in countries where research funding is scarce.
Article
Accurate detection of individual animals is integral to the management of vulnerable wildlife species, but often difficult and costly to achieve for species that occur over wide or inaccessible areas or engage in cryptic behaviours. There is a growing acceptance of the use of drones (also known as unmanned aerial vehicles, UAVs and remotely piloted aircraft systems, RPAS) to detect wildlife, largely because of the capacity for drones to rapidly cover large areas compared to ground survey methods. While drones can aid the capture of large amounts of imagery, detection requires either manual evaluation of the imagery or automated detection using machine learning algorithms. While manual evaluation of drone‐acquired imagery is possible and sometimes necessary, the powerful combination of drones with automated detection of wildlife in this imagery is much faster and, in some cases, more accurate than using human observers. Despite the great potential of this emerging approach, most attention to date has been paid to the development of algorithms, and little is kn