
Multi-Object Tracking with Memory Fusion in UAV Videos
Author(s) -
Yibo Cui,
Shangsheng Li,
Xin Yang,
Gang Wang
Publication year - 2025
Publication title -
ieee access
Language(s) - English
Resource type - Magazines
SCImago Journal Rank - 0.587
H-Index - 127
eISSN - 2169-3536
DOI - 10.1109/access.2025.3596684
Subject(s) - aerospace , bioengineering , communication, networking and broadcast technologies , components, circuits, devices and systems , computing and processing , engineered materials, dielectrics and plasmas , engineering profession , fields, waves and electromagnetics , general topics for engineers , geoscience , nuclear engineering , photonics and electrooptics , power, energy and industry applications , robotics and control systems , signal processing and analysis , transportation
Multi-object tracking (MOT) plays a pivotal role in numerous UAV-related tasks. Nevertheless, conventional approaches often encounter limitations when facing challenges such as motion blur and target deformation, primarily due to their dependence on local features and static spatial representations. To overcome these constraints, we proposeAMF-MOT, an innovative framework featuring an Adaptive Memory Fusion module that exploits rich spatio-temporal information. Our method centers around a specialized short-term memory structure that adaptively retrieves relevant information through an attention mechanism and efficiently fuses multi-frame features via a dedicated fusion module. This design enables robust multi-frame dependency modeling and efficient memory propagation, thereby improving object association and re-identification performance. The AMF module surpasses existing methods by offering key advantages: lightweight, plug-and-play, and features fixed computational complexity without requiring a predefined number of input frames. we achieved an Identification F1 Score (IDF1) of 52.8% and a Multiple Object Tracking Accuracy (MOTA) of 41.2% on the VisDrone2019 dataset, and achieved an IDF1 of 69.2% and a MOTA of 48.8% on the UAVDT dataset. The model operates in real-time, making it suitable for time-critical UAVapplications. In-depth ablation studies further validate the effectiveness of theAMFmodule particularly in challenging scenarios involving occlusions and motion blur. In this paper, we contribute a novel memory fusion mechanism, a lightweight MOT architecture, and improved ID association performance by using the AMF module. The source code will be publicly available at: https://github.com/keacifer/AMF-MOT.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom