z-logo
open-access-imgOpen Access
From Blurs to Birds: Localization and Classification of Hard-to-See Bird Species in Norwegian Wilderness Camera Trap Images
Author(s) -
Havard Teigen,
Ammar Ahmed,
Ali Shariq Imran,
Mohib Ullah,
R. Muhammad Atif Azad,
Ahmet Soylu
Publication year - 2025
Publication title -
ieee access
Language(s) - English
Resource type - Magazines
SCImago Journal Rank - 0.587
H-Index - 127
eISSN - 2169-3536
DOI - 10.1109/access.2025.3613068
Subject(s) - aerospace , bioengineering , communication, networking and broadcast technologies , components, circuits, devices and systems , computing and processing , engineered materials, dielectrics and plasmas , engineering profession , fields, waves and electromagnetics , general topics for engineers , geoscience , nuclear engineering , photonics and electrooptics , power, energy and industry applications , robotics and control systems , signal processing and analysis , transportation
Camera trap imagery datasets present unique challenges, including weather conditions, motion blur, occlusions from vegetation or other animals, and instances where the animal is partially or completely outside the frame due to the camera’s fixed position. This challenge is exacerbated when the target animal species are very small, highly similar in appearance, and fast-moving, such as different species of birds. Although deep learning has been extensively applied to classify and detect animal species, bird species are either captured in static, high-quality images under optimal conditions, with no motion blur or other disturbances, or they are grouped into a single class. Real camera trap images, however, are not optimal nor of high quality. Furthermore, recognizing bird species is more detailed and intricate than identifying larger animals such as deer, moose, or reindeer, making manual annotation more time-consuming and susceptible to errors. This study utilizes state-of-the-art deep-learning neural networks to localize and classify hard-to-see bird species in Norwegian wilderness camera trap images. Experimental results show that our fine-tuned YOLOv8x architecture achieved the highest performance for localization, outperforming YOLOv9, yielding a mAP@50 score of 0.92. For the classification of hard-to-distinguish bird species, EfficientNetB7 achieved the highest accuracy and F1-score of 0.87. Grad-CAM, an explainable AI technique, was employed to identify discriminative regions within the images.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom