Sitemap

A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.

Pages

Posts

Future Blog Post

less than 1 minute read

Published:

This post will show up by default. To disable scheduling of future posts, edit config.yml and set future: false.

Blog Post number 4

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 3

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 2

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 1

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

portfolio

publications

An Efficient Detection Framework for Aerial Imagery Based on Uniform Slicing Window

Published in Remote Sensing, 2023

Drone object detection faces numerous challenges such as dense clusters with overlapping, scale diversity, and long-tail distributions. Utilizing tiling inference through uniform sliding window is an effective way of enlarging tiny objects and meanwhile efficient for real-world applications. However, merely partitioning input images may result in heavy truncation and an unexpected performance drop in large objects. Therefore, in this work, we strive to develop an improved tiling detection framework with both competitive performance and high efficiency. First, we formulate the tiling inference and training pipeline with a mixed data strategy. To avoid truncation and handle objects at all scales, we simultaneously perform global detection on the original image and local detection on corresponding sub-patches, employing appropriate patch settings. Correspondingly, the training data includes both original images and the patches generated by random online anchor-cropping, which can ensure the effectiveness of patches and enrich the image scenarios. Furthermore, a scale filtering mechanism is applied to assign objects at diverse scales to global and local detection tasks to keep the scale invariance of a detector and obtain optimal fused predictions. As most of the additional operations are performed in parallel, the tiling inference remains highly efficient. Additionally, we devise two augmentations customized for tiling detection to effectively increase valid annotations, which can generate more challenging drone scenarios and simulate the practical cluster with overlapping, especially for rare categories. Comprehensive experiments on both public drone benchmarks and our customized real-world images demonstrate that, in comparison to other drone detection frameworks, the proposed tiling framework can significantly improve the performance of general detectors in drone scenarios with lower additional computational costs.

Recommended citation: Yang, X.; Song, Y.; Zhou, Y.; Liao, Y.; Yang, J.; Huang, J.; Huang, Y.; Bai, Y. An Efficient Detection Framework for Aerial Imagery Based on Uniform Slicing Window. Remote Sens. 2023, 15, 4122.
Download Paper

Multispectral Sample Augmentation and Illumination Guidance for RGB-T Object Detection by MMDetection Framework

Published in 2024 International Conference on Computer Network and Cloud Computing (CNCC 2024), 2024

Multispectral object detection technology has important application prospects in the fields of autonomous driving and so on. Conventional multispectral object detection algorithm rely solely on deep neural networks to learn multispectral image sample information, lacking the guidance of prior knowledge, and not fully utilizing infrared, visible, and other spectral information, resulting in decreased accuracy of object detection in complex scenes. To address this problem, this paper proposes an object detection algorithm based on infrared visible sample augmentation and illumination guidance. The algorithm adopts the MMDetection framework and extracts multispectral object features based on a designed sample augmentation method based on the fusion of positive and negative samples in multispectral images. Based on a designed adaptive weight allocation method guided by illumination, it enhances the algorithm’s adaptability to the lighting environment. Finally, through the design of a multi-task loss function, it achieves high-precision and robust object detection in complex scenes.

Recommended citation: Jinqi Yang, Xin Yang, Yizhao Liao, Jinxiang Huang, Hongyu He, Erfan Zhang, Ya Zhou, Yong Song*. Multispectral Sample Augmentation and Illumination Guidance for RGB-T Object Detection by MMDetection Framework[C]. 2024 4th International Conference on Laser, Optics and Optoelectronic Technology.
Download Paper

Light Siamese Network for Long-Term Onboard Aerial Tracking

Published in IEEE Transactions on Geoscience and Remote Sensing, 2024

The scarce onboard computational resources and real-time demand restrict the deployment of aerial trackers with sophisticated structures and customized operators. Meanwhile, aerial trackers need updating modules to adapt to continuous appearance variations in real-world long-term (LT) scenarios. However, frequent updating will introduce noisy templates and lead to tracking drifts and efficiency drops. Therefore, in this work, we develop a lightweight and highly efficient Siamese tracker for LT onboard aerial tracking applications. First, we build a compact, plain, and deployment-friendly Siamese network based on re-parameterization (Rep) as the baseline short-term (ST) tracker. Furthermore, we propose a tracking-specific decoupled knowledge distillation (KD) guided by strict teachers to unleash the appearance representation potential of the feature extractor without extra inference cost. Specifically, before distillation, the teacher conducts qualification verification to avoid misguiding the student. Then, hard negative background regions are mined and decoupled with the target region, encouraging the student to focus more on similar distractors and informative areas. Finally, to realize efficient and high-confidence LT tracking, we design two extensions and incorporate them into the boosted ST tracker: an initial-template-driven template updater with a corresponding pair-generating strategy to alleviate appearance pollution, and a confidence estimating branch to determine whether to update. Extensive results on large-scale drone benchmarks indicate that our proposed tracker significantly outperforms state-of-the-art (SOTA) aerial trackers. Real-world tests on our customized drone-captured LT dataset also validate its favorable practicability with a real-time speed of 44 fps on the Lynix KA200 chip.

Recommended citation: X. Yang, J. Huang, Y. Liao, Y. Song, Y. Zhou and J. Yang, "Light Siamese Network for Long-Term Onboard Aerial Tracking," in IEEE Transactions on Geoscience and Remote Sensing, vol. 62, pp. 1-15, 2024.
Download Paper

RDR-KD: A Knowledge Distillation Detection Framework for Drone Scenes

Published in IEEE Geoscience and Remote Sensing Letters, 2024

Drone object detection (DOD) with real-time deployment is a research hotspot. On the one hand, the performance of tiny object detection is closely related to the ground detection capability of the drone platform. Existing methods are keen on designing complex networks to enhance the accuracy of tiny objects, which significantly increases computational costs. On the other hand, the limited drone hardware resources urgently require lightweight models for deployment. To address the dilemma of balancing detection accuracy and computational efficiency, we propose a regenerated-decoupled-responsive knowledge distillation (RDR-KD) framework specifically for drone scenes. First, we design the Regenerated Distillation and the Decoupled Distillation to fully transfer the tiny object feature information from the teacher model to the student model. Meanwhile, we devise the logit-based Responsive Distillation based on focal loss and efficient intersection over union (EIoU) to alleviate class imbalance. Finally, we conduct extensive experiments on the VisDrone2019 dataset. The experimental results demonstrate that the proposed RDR-KD framework improves AP and AP_S of the student model by 3.3% and 2.9% respectively, which outperforms other state-of-the-art distillation frameworks.

Recommended citation: J. Huang, H. Chang, X. Yang, Y. Liu, S. Liu and Y. Song, "RDR-KD: A Knowledge Distillation Detection Framework for Drone Scenes," in IEEE Geoscience and Remote Sensing Letters, vol. 21, pp. 1-5, 2024
Download Paper

talks

teaching

Teaching experience 1

Undergraduate course, University 1, Department, 2014

This is a description of a teaching experience. You can use markdown like any other post.

Teaching experience 2

Workshop, University 1, Department, 2015

This is a description of a teaching experience. You can use markdown like any other post.