337
Views
2
CrossRef citations to date
0
Altmetric
Research Article

Lightweight detection network for arbitrary-oriented vehicles in UAV imagery via precise positional information encoding and bidirectional feature fusion

ORCID Icon, ORCID Icon & ORCID Icon
Pages 4529-4558 | Received 26 Sep 2022, Accepted 26 Mar 2023, Published online: 19 Apr 2023
 

ABSTRACT

Unmanned aerial vehicles (UAVs) open up new opportunities for transportation monitoring. However, the vehicle targets in UAV images are situated in dynamic scenarios, such as uncertain background, dramatically varying arrangement density, multi-scale, and arbitrary-oriented. Most strategies for UAV-based monitoring require complex manoeuvring and still lack accurate abilities and lightweight structures. Consequently, designing effective detection methods with both speed and accuracy is challenging. This paper proposes a lightweight YOLO-based arbitrary-oriented vehicle detector via precise positional information encoding and bidirectional feature fusion to address the above issues. First, an additional angular classification prediction branch is added to the YOLO head network to significantly improve the detection performance for arbitrary-oriented vehicles without incurring the extra computational complexity and burden. Second, a C3 module embedded coordinate attention (C3CA) is presented to capture long-range dependencies and preserve vehicles’ precise positional information in feature maps. Then, a fully connected bidirectional feature fusion module (FC-BiFPN) is applied at the neck of the YOLO detection framework, which is helpful for multi-scale vehicle detection. This module can efficiently aggregate features at different resolutions and automatically enhance information interaction. Finally, experiments and comparisons on vehicle and remote sensing datasets demonstrate that our approach outperforms the state-of-the-art methods in balancing precision and efficiency. In addition, the overall network design follows the lightweight concept, which better meets the real-time requirements of the UAV urban traffic monitoring platform in realistic scenarios.

Disclosure statement

No potential conflict of interest was reported by the authors.

Data availability statement

UAV-ROD dataset available at https://github.com/fengkaibit/UAV-ROD.

UCAS-AOD dataset available at https://hyper.ai/datasets/5419.

Source code available at https://github.com/2876434806/YOLO_rotate.

Additional information

Funding

The work was supported by the National Natural Science Foundation of China (41971365).

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.