Next Article in Journal
On Intersections of B-Spline Curves
Next Article in Special Issue
Multi-Camera Multi-Vehicle Tracking Guided by Highway Overlapping FoVs
Previous Article in Journal
New Results on the Ulam–Hyers–Mittag–Leffler Stability of Caputo Fractional-Order Delay Differential Equations
Previous Article in Special Issue
High-Efficiency and High-Precision Ship Detection Algorithm Based on Improved YOLOv8n
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Oriented Object Detection by Scene Classification and Task-Aligned Focal Loss

College of Electrical and Information Engineering, Zhengzhou University of Light Industry, Zhengzhou 450002, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(9), 1343; https://doi.org/10.3390/math12091343
Submission received: 25 March 2024 / Revised: 17 April 2024 / Accepted: 25 April 2024 / Published: 28 April 2024
(This article belongs to the Special Issue Advances in Computer Vision and Machine Learning, 2nd Edition)

Abstract

:
Oriented object detection (OOD) can precisely detect objects with arbitrary direction in remote sensing images (RSIs). Up to now, the two-stage OOD methods have attracted more attention because of their high detection accuracy. However, the two-stage methods only rely on the features of each proposal for object recognition, which leads to the misclassification problem because of the intra-class diversity, inter-class similarity and clutter backgrounds in RSIs. To address the above problem, an OOD model combining scene classification is proposed. Considering the fact that each foreground object has a strong contextual relationship with the scene of the RSI, a scene classification branch is added to the baseline OOD model, and the scene classification result of input RSI is used to exclude the impossible categories. To focus on the hard instances and enhance the consistency between classification and regression, a task-aligned focal loss (TFL) which combines the classification difficulty with the regression loss is proposed, and TFL assigns lager weights to the hard instances and optimizes the classification and regression branches simultaneously. The ablation study proves the effectiveness of scene classification branch, TFL and their combination. The comparisons with 15 and 14 OOD methods on the DOTA and DIOR-R datasets validate the superiority of our method.

1. Introduction

Oriented object detection (OOD) has a significant role in remote sensing image (RSI) interpretation, which can localize objects with arbitrary direction more accurately by using an oriented rectangular box, comparing with horizontal object detection [1,2,3]. OOD can be applied in resource exploration, urban planning, modern agriculture, military reconnaissance and so on [4,5,6,7,8,9,10].
The existing OOD methods can be classified into three categories, including anchor-free methods, one-stage anchor-based methods and two-stage anchor-based methods. Anchor-free methods directly predict the bounding box and category score for each feature point through fully convolutional networks, such as the dynamic refinement network (DRN) [11], oriented representative points (Oriented RepPoints) [12], etc. One-stage methods predefine a number of anchors for each feature point and then use the fully convolutional network to predict the localization offset and category score of each predefined anchor, such as the single-shot alignment network ( S 2 A-Net) [13], dynamic anchor learning (DAL) [14], etc. [15,16,17,18,19,20,21,22]. Two-stage methods add a region proposal network (RPN), which predicts object proposals from a large number of predefined anchors, and then the localization offset and category score of each proposal are predicted according to its features obtained through a rotated region of interest alignment (RRoIAlign) operation, such as Oriented R-CNN [23], anchor-free oriented proposal generator (AOPG) [24], etc. [25,26,27].
On the whole, the detection accuracy of two-stage methods is better than the other two types of method; therefore, most works have focused on the two-stage methods. However, they suffer from a common problem, i.e., the misclassification caused by solely relying on the features of each proposal itself. Compared with natural scene images, the intra-class diversity, inter-class similarity and complexity of the background are more severe in remote sensing images. In this situation, solely using the features of proposal can easily lead to the misclassification problem. As shown in Figure 1a, a background region is incorrectly identified as a harbor.
To address the above problem, in this paper an OOD model that fuses scene classification is proposed. As shown in Figure 2, a scene classification branch is added to the baseline OOD model, and the result of scene classification can help the object classification branch to exclude the impossible classes. As a matter of fact, each foreground object has a strong contextual relationship with the scene of RSI; consequently, making full use of scene classification to assist object recognition can effectively alleviate the aforementioned misclassification problem. As shown in Figure 1b, our method can effectively exclude the misclassification of the harbor with the help of scene classification.
Furthermore, to focus on the hard instances and enhance the consistency between classification and regression, in this paper a novel task-aligned focal loss (TFL) is proposed. Specifically, the classification difficulty of each instance is combined with the regression loss to obtain the TFL, which can increase the relative weights of hard instances and optimize the classification and regression of each instance simultaneously.
The major contributions of this paper can be summarized as follows:
  • A novel OOD model fusing scene classification is proposed to address the misclassification problem caused by solely relying on the features of each proposal itself. Owing to intra-class diversity and inter-class similarity, the misclassification of objects easily occurs if only the features of each proposal itself are used for object classification. Considering the contextual relationship between the foreground objects and the scene of an RSI, the scene classification branch is incorporated into the baseline OOD model help the object classification branch exclude impossible categories that do not exist in the RSI;
  • A novel TFL is proposed to focus on the hard instances and enhance the consistency between classification and regression. The TFL is obtained through combining the classification difficulty with regression loss, which can increase the relative weights of hard instances and achieve the simultaneous optimization of the classification and regression branches;
  • The superiority of the proposed method is demonstrated on two public RSI datasets, i.e., the DOTA and DIOR-R benchmarks.

2. Related Work

2.1. Oriented Object Detection Methods

Due to the development of deep learning [28,29,30,31,32,33,34,35,36,37,38,39,40,41,42], the OOD of RSI has made rapid progress in recent years [43,44,45,46,47,48,49,50,51,52,53]. The current OOD methods are generally divided into three categories, i.e., anchor-free methods, one-stage anchor-based methods and two-stage anchor-based methods, and their representative approaches are briefly introduced as follows.
Anchor-free methods: DRN; in [11], the authors proposed a feature selection module that allows the receptive fields of neurons to dynamically adapt to the shape and orientation of objects. Moreover, a dynamic refinement head is proposed to model the uniqueness and specificity of each sample by means of object-awareness to refine the features and thus make better inferences. Oriented RepPoints; in [12], the authors utilized adaptive point sets to represent the oriented boxes and designed an adaptive points assessment assignment strategy to select typical samples for better training.
One-stage anchor-based methods: S 2 A-Net; in [13], the authors proposed an aligned convolution module to produce high-quality proposals that could relieve the inconsistency problem between classification and regression to some extent. DAL; in [14], the authors proposed a dynamic anchor learning strategy that combined spatial alignment with feature alignment capabilities for better label assignment. Refined single-stage detector (R3Det); in [54], the authors proposed a feature refinement module and a SkewIou loss to address the detection of objects with large aspect ratios and dense distributions in arbitrary directions.
Two-stage anchor-based methods: RoI Transformer; in [55], the authors converted horizontal proposals to oriented proposals through the proposed RoI learner for accurate OOD. Oriented R-CNN; in [23], the authors proposed an oriented RPN that simply generated high-quality proposals by designing an innovative oriented proposals representation. AOPG; in [24], the authors proposed a coarse location module to acquire oriented anchors and refine them to obtain high-quality oriented proposals. Dual-aligned oriented detector (DoDet); in [56], the authors designed a localization-guided detection head to mitigate the problem of inconsistency between classification and regression. On improving bounding box representations (OIBBR); in [57], the authors proposed a quadrant point representation to handle the boundary discontinuity problem.

2.2. RoI Pooling Methods

For the two-stage detection model, the region of interest (RoI) pooling operation is imposed on the feature maps of each proposal to obtain a feature representation with fixed size. For the original RoI pooling method [58], the feature maps of each proposal are approximately divided into 7 × 7 bins in the spatial dimension, and the max pooing operation is applied to obtain a 7 × 7 feature map, which is used as a channel of feature maps of each proposal. However, the original method suffers from quantization errors, i.e., the size of each proposal is not always evenly divisible by 7.
To address the above problem, an RoI Alignment (RoIAlign) operation [59] is proposed to replace original RoI pooling operation, and it is briefly introduced as follows. Firstly, the feature maps of each proposal are exactly divided into 7 × 7 bins in the spatial dimension, even if the size of the proposal is not evenly divisible by 7; in other words, the coordinates of each bin are decimals. Afterwards, each bin is exactly divided into four regions, and the pixel value of the center point of each region is calculated through the bilinear interpolation of four adjacent pixels of the center point. Finally, the value of each bin is obtained by taking the maximum value of the four center points, and the 7 × 7 feature representation of each proposal in the spatial dimension is obtained without the quantization error.
Note that the RoIAlign operation is designed for horizontal proposals, and cannot be directly applied to oriented proposals. Therefore, an RRoIAlign operation [55] is proposed to handle above restriction, which is a rotated version of RoIAlign operation, i.e., the only difference of them is rotation. So far, the RRoIAlign operation has been widely used in almost all OOD methods besides our method. To the best of our knowledge, no better RoI pooling method has been proposed to replace the RRoIAlign operation.

3. Proposed Method

3.1. Overview

The architecture of the proposed method is illustrated in Figure 2. First of all, the AOPG [24] is used to generate high-quality oriented proposals. Secondly, the classification and regression results of each proposal are derived from its features, obtained through the RROI Align operation, and the classification and regression branches are trained by the proposed TFL. Thirdly, a scene classification branch is added on the basis of shared backbone features, and it is trained by the image-level labels derived from the instance-level labels. In the inference stage, the predicted results of the scene classification branch are binarized by a predefined threshold, and they are then used to exclude the impossible categories for object classification. Finally, the refined category scores and predicted offsets of all proposals are jointly used to determine the final detection results. The proposed scene classification branch and TFL are introduced in detail as follows.

3.2. Scene Classification Branch

As mentioned previously, the existing two-stage methods usually suffer from the misclassification problem caused by solely relying on the features of each proposal itself. Considering the fact that each foreground object has a strong contextual relationship with the scene of the RSI, a scene classification branch is added into the OOD model to assist the object classification.
As shown in Figure 2, the scene classification branch is constructed on the basis of a shared backbone. First of all, the feature maps output from the shared backbone are converted into a one-dimensional feature vector by using the global average pooling operation, which are irrelevant with the features obtained from the RRoI Align operation. Secondly, the feature vector is convolved by a fully connected layer to obtain the scene feature vector, denoted as f R C + 1 , where C indicates the number of categories. Finally, the scene classification result of input RSI, denoted as s, can be obtained through the following equation:
s i = S i g m o i d ( f i ) , s . t . i Z , 1 i C + 1 ,
where f i indicates the ith element in f, s i indicates the probability that input RSI belongs to the ith category, the ( C + 1 ) th category denotes background and S i g m o i d ( · ) denotes the sigmoid activation function. It is worth noting that the softmax classifier is not suitable for the scene classification branch because a multi-classification task is involved here; consequently, as shown in Figure 2, the sigmoid function is adopted as a binary classifier for the classification of each category.
At this point, the training loss of the scene classification branch, denoted as L S C B , is defined by using the binary cross entropy (BCE) loss:
L S C B = i = 1 C + 1 l i log s i + ( 1 l i ) log ( 1 s i ) ,
where l i indicates the image-level label of the ith category of input RSI; l i = 1 if input RSI contains the objects belonging to ith category, otherwise, l i = 0 . As mentioned previously, the classifier of the scene classification branch consists of multiple binary classifiers; therefore, the BCE loss of each category is accumulated to obtain the L S C B .

3.3. Task-Aligned Focal Loss

The focal loss [60] has demonstrated that paying more attention to the hard instances can effectively enhance the detection capability; however, it cannot handle the inconsistency between the classification and regression. To focus on the hard instances and enhance the consistency between classification and regression simultaneously, a task-aligned focal loss, denoted as L T F L , is proposed, and its formulation is given as follows:
L T F L = L c l s + α L r e g
L c l s = log ( p i ) , if y i = 1 and 1 i C ( 1 p i ) log ( 1 p i ) , if y i = 1 and i = C + 1
L r e g = e ( 1 p i ) L R I o U , s . t . y i = 1 and 1 i C ,
where L c l s and L r e g denote the classification and regression loss in L T F L , respectively, α denotes the relative weight of L r e g and is quantitatively analyzed in Section 4.2, p i p indicates the probability that each proposal belongs to the ith category, p C + 1 indicates the classification result of each proposal, y i denotes the instance-level label of the ith category of each proposal, α is a predefined hyper-parameter and L R I o U denotes the RIoU loss [50], which proposed a new metric named rotated intersection of over union (RIoU) and optimized the RIoU for better bounding box regression (BBR).
The explanation of L T F L is as follows. Each instance will be used for classification and regression if it belongs to the assembly of positive samples, i.e., y i = 1 and 1 i C . At this point, the classification difficulty of each instance, indicated by e ( 1 p i ) , is adaptively adjusted with p i and is used as the weight of the L R I O U . On the one hand, the hard instances will be assigned larger weights compared with the easy instances in L r e g . On the other hand, the L R I O U is optimized and the p i is converged to 1 simultaneously, along with the minimization of L r e g ; in other words, the consistency between classification and regression is enhanced. At this point, the traditional cross entropy, indicated by log ( p i ) , is used for the training of classification of positive samples. The proposal will only be used for classification if it belongs to the assembly of negative samples, i.e., y i = 1 and i = C + 1 ; consequently, the original focal loss, indicated by ( 1 p i ) log ( 1 p i ) , is used for the training of the classification of negative samples.

3.4. Overall Training Loss

The overall training loss of our OOD method, denoted as L, is formulated as follows:
L = L R P N + L S C B + L T F L
L R P N = L c l s + L r e g ,
where L R P N denotes the loss of the RPN, L c l s denotes the training loss of the classification branch in RPN and is defined by using cross entropy, L r e g denotes the training loss of the BBR branch in RPN and is defined by using the smooth L1 loss [58], and the details of L R P N can be seen in the baseline OOD model [24].

3.5. Inference Stage

First of all, the scene classification result of the input RSI is predicted by the trained scene classification branch, and it is binarized through the following equation:
t s i = 1 , if s i T 0 , otherwise , t s i t s ,
where t s R C + 1 denotes the binarized scene classification result of input RSI, t s i denotes the result of the ith category in t s , T is a predefined threshold and is quantitatively analyzed in Section 4.2. In other words, t s i = 0 if the input RSI does not contain the objects belonging to the ith category, otherwise, t s i = 1 . Consequently, the t s can be used to exclude impossible categories for each proposal:
p s i = p i × t s i , p s i p s ,
where p s R C + 1 denotes the final classification result of each proposal, p s i denotes the result of the ith category in p s . The motivation of Equation (9) is as follows. The possibility that a proposal belongs to the ith category will be excluded if the predicted result of the scene classification branch indicates that the input RSI does not contain the scene of the ith category. Finally, the p s is used for non-maximumsuppression (NMS) [61].

4. Experiments

4.1. Experiment Setup

4.1.1. Datasets

The DOTA dataset [62] includes 2806 images and 188,282 instances with 15 categories, i.e., plane (PL), baseball (BD), bridge (BR), ground track and field (GTF), small vehicle (SV), large vehicle (LV), ship (SH), tennis court (TC), basketball court (BC), storage tank (ST), soccer fields (SBF), rings (RA), harbors (HA), swimming pools (SP) and helicopters (HC). For single-scale experiments, the raw RSI is cropped into several sub-images with 1024 × 1024 pixels by using a step with 824 pixels. For multi-scale experiments, the raw RSI is initially resized with three ratios, i.e., 0.5, 1 and 1.5, and then the resized RSIs are cut into multiple sub-images with 1024 × 1024 pixels by using a step with 524 pixels. All experimental results were obtained through the DOTA evaluation server.
The DIOR-R dataset [63] includes 23,463 images with 800 × 800 pixels and 192,518 instances with 20 categories, i.e., aircraft (APL), airport (APO), baseball field (BF), BC, BR, chimney (CH), highway service area (ESA), expressway toll station (ETS), dam (DAM), golf field (GF), GTF, HA, overpass (OP), SH, stadium (STA), storage tank (STO), tennis court (TC), train station (TS), vehicle (VE) and windmill (WM).
Similar to other OOD methods [24,49,50,56,57], vertical and horizontal flipping are also used to augment the training samples for fair comparison.

4.1.2. Implementation Details

The stochastic gradient descent (SGD) algorithm was used to optimize our model with a momentum of 0.9 and a weight decay of 0.0001. The initial learning rate, batchsize and the number of epoch were set to 0.02, 8 and 12, respectively. The learning rate was modified to the 0.1-time previous stage at the eighth and eleventh epochs. The threshold of NMS is 0.1 (0.5) on the DOTA (DIOR-R) dataset [62,63].
The experiments were conducted on the mmdetection platform [64] within the PyTorch framework and run on four NVIDIA GeForce RTX 2080Ti (4×11-GB memory).
The mean average precision (mAP) of each category and all categories was used to assess the performance of the proposed method.

4.2. Parameter Analysis

The α (T) in Equation (3) (Equation (8)) is quantitatively analyzed on the DOTA dataset to assess its impact on the final result. As shown in Figure 3 and Figure 4, the highest mAP are obtained at α = 0.5 ( T = 0.1 ); therefore, in this paper the α (T) is set to 0.5 (0.1). As shown in Equation (3), the α is used to control the relative weight of classification loss and BBR loss; therefore, a too-large or too-small value is not suitable for α .

4.3. Ablation Study

Ablation studies were conducted on the DOTA dataset to evaluate the contribution of the scene classification branch and task-aligned focal loss. The AOPG [24] was applied as the benchmark methodology. As shown in Table 1, the baseline + SCB (TFL) denotes the combination of the baseline method and scene classification branch (task-aligned focal loss), and the baseline +SCB + TFL denotes our method in which the scene classification branch and task-aligned focal loss are all used.
As shown in Table 1, compared with baseline, the mAP of baseline + SCB, baseline + TFL and baseline +SCB + TFL are increased by 1.51%, 1.45% and 1.77%, respectively, which verifies the effectiveness of SCB, TFL and their combination.

4.4. Comparisons with Other OOD Methods

To fully validate the capability of our method (denoted as SCTFL), it is compared with several other OOD models on two RSI benchmarks, and the details are as follows.

4.4.1. Results on the DOTA Dataset

As shown in Table 2, we compare our model with 15 other OOD methods on the DOTA dataset: RetinaNet-O [60]; Faster RCNN [31]; DRN [11], CenterMap-Net [65]; multi-category rotation detector for small, cluttered and rotated objects (SCRDet) [66]; S 2 A-Net [13]; dynamic prior along with the coarse-to-fine assigner (DCFL) [67]; ROI Transformer [55]; task-collaborated detector (TCD) [68]; AOPG [24]; DODet [56]; Oriented R-CNN [23]; Oriented RepPoints [12]; OIBBR [57] and rotation proposal generation and optimization detector (RPGAOD) [69], and the mAP of our method is boosted by 8.52%, 7.90%, 6.25%, 5.21%, 4.34%, 2.83%, 2.69%, 2.34%, 1.77%, 1.71%, 1.46%, 1.08%, 0.98%, 0.70% and 0.48%, respectively, which shows our best performance on the DOTA dataset.

4.4.2. Results on the DIOR-R Dataset

As shown in Table 3, we compare our method with 14 other OOD methods on the DIOR-R dataset: fully convolutional one-stage object detection (FCOS-O) [70], RetinaNet-O [60], Double-heads [71], Faster RCNN-O [31], Gliding Vertex [65], S 2 A-Net [13], ROI Transformer [55], OIBBR [57], AOPG [24], Oriented R-CNN [23], TCD [68], DODet [56], Oriented RepPoints [12] and DCFL [67], and the mAP of our method is boosted by 15.89%, 9.73%, 7.83%, 7.74%, 7.22%, 4.38%, 3.41%, 3.08%, 2.87%, 2.65%, 2.24%, 2.18%, 0.57% and 0.48%, respectively, which shows our best performance on the DIOR-R dataset.

4.5. Evaluation of Generalizability

4.5.1. Evaluation under the Condition of Multi-Scale Training and Testing

To evaluate the generalizability of our method through varying spatial resolutions and object scales, seven OOD methods are compared with our method under the condition of multi-scale training and testing: DRN [11], S 2 A-Net [13], TCD [68], AOPG [24], DODet [56], Oriented R-CNN [23] and OIBBR [57]. These provide multi-scale experiment results. For multi-scale training and testing, the raw RSI is initially resized with three ratios, i.e., 0.5, 1 and 1.5, and then the resized RSIs are cut into multiple sub-images with 1024 × 1024 pixels by using a step with 524 pixels. As shown in Table 4, our method achieves the best results among the eight OOD methods, which validates the generalizability of our method under the condition of multi-scale training and testing.

4.5.2. Evaluation in Densely Populated Scenes

To evaluate the generalizability of our method in densely populated scenes, it is compared with popular OOD methods in the four densely populated scenes of two RSI datasets, respectively. The four densely populated scenes of the DOTA dataset include SV, LV, SH and TC, and scenes of the DIOR-R dataset include APL, SH, TC and VE. As shown in Table 5 and Table 6, our method has the highest mAP in the four densely populated scenes of the two RSI datasets, which indicates that our method has better generalizability in densely populated scenes.
In addition, as shown in Table 2, Table 3 and Table 4, the overall performance of our method is the best in 29 object categories of the DOTA and DIOR-R datasets (a total of 35 categories eliminate 6 duplicated categories), which demonstrates that our method has excellent generalizability in different object categories, and it can validate certain matters to some extent, i.e., our method can provide good performance when novel object categories are introduced.

4.6. Subjective Evaluation

To intuitively evaluate the effectiveness of our method, some its detection results on the two benchmarks are visualized in Figure 5 and Figure 6, respectively.

5. Discussion

5.1. Analysis of Experiment Results

As shown in Table 2 (Table 3), our method achieved the best performances in 4 (9) categories, sub-optimal performances in 2 (3) categories and the third best performances in 1 (2) categories on the DOTA (DIOR-R) datasets, respectively. This is attributed to the effectiveness of the proposed scene classification branch and task-aligned focal loss. For example, as shown in Table 3, our method gives the best performance in the ship category. A possible reason for this is that our method can exclude the misclassification of ships with the help of the scene classification branch since ships are closely related to a certain scene, e.g., sea. For another example, most of the OOD methods give poor performance in bridge and vehicle categories; however, our method shows the best performance in the above two categories. A possible reason for this is that the proposed task-aligned focal loss can assign larger weights to hard instances and enhance the consistency of classification and regression tasks. Furthermore, as shown in Table 2 and Table 3, the mAP of all categories of our method is the highest among all OOD methods, which also demonstrates the effectiveness of the scene classification branch and task-aligned focal loss.

5.2. Analysis of Failure Results

Although the proposed method provides better performance compared with other OOD methods, it still obtains failure results on some occasions, and the reasons are summarized below.
  • Caused by inter-class similarity. It is difficult for OOD methods to distinguish objects with similar characteristics but are of different categories because of the inter-class similarity of RSI, and our method also cannot handle it if similar objects exist in one RSI simultaneously. As a matter of fact, the scene classification branch introduced by our method can only exclude the impossible categories that do not exist in the input RSI. As shown in Figure 7a, the large vehicles are mistakenly detected as small vehicles because of their similar characteristics.
  • Caused by shadows. The shadows of foreground objects have similar shapes to the foreground objects themselves; therefore, on some occasions they are mistakenly detected as a part of the foreground objects. As shown in Figure 7b, the shadows of a chimney and a windmill are mistakenly recognized as a part of the targets.

6. Conclusions

This paper proposes an OOD model combining scene classification to handle the misclassification problem caused by only relying on the features of each proposal itself. The scene classification branch is added into the baseline OOD model, and the scene classification results are employed to help the object classification branch to exclude impossible categories in the RSI. In addition, a task-aligned focal loss is proposed to focus on hard instances and enhance the consistency between classification and regression, which combines the instance difficulty with regression loss to increase the relative weight of hard instance and optimize the classification and regression branches simultaneously. Ablation experiments show the effectiveness of the scene classification branch, task-aligned focal loss and their fusion. Comparisons with 15 and 14 OOD methods on the DOTA and DIOR-R datasets demonstrate the excellent performance of our method.

Author Contributions

Conceptualization, X.Q.; formal analysis, X.Q.; funding acquisition, X.Q.; methodology, X.Q. and S.G.; project administration, W.W.; resources, W.W.; software, S.G.; supervision, W.W.; validation, W.W. and W.D.; writing—original draft, S.G.; writing—review and editing, X.Q. and W.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (Grant No. 62076223), the Key Research Project of Henan Province Universities (Grant No. 24ZX005) and the Key Science and Technology Program of Henan Province (Grant No. 232102211018).

Data Availability Statement

The DOTA and DIOR-R datasets are available at the following URLs: https://captain-whu.github.io/DOTA/index.html (accessed on 15 March 2023) and https://gcheng-nwpu.github.io/ (accessed on 15 March 2023), respectively.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
OODOriented Object Detection
RSIRemote Sensing Image
SCBScene Classification Branch
TFLTask-aligned Focal Loss

References

  1. Yao, X.; Shen, H.; Feng, X.; Cheng, G.; Han, J. R2IPoints: Pursuing Rotation-Insensitive Point Representation for Aerial Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5517010. [Google Scholar] [CrossRef]
  2. Wu, X.; Hong, D.; Chanussot, J. Convolutional Neural Networks for Multimodal Remote Sensing Data Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5517010. [Google Scholar] [CrossRef]
  3. Hong, D.; Yokoya, N.; Chanussot, J.; Zhu, X.X. An Augmented Linear Mixing Model to Address Spectral Variability for Hyperspectral Unmixing. IEEE Trans. Image Process. 2019, 28, 1923–1938. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, L.; Zhang, Y.; Yan, H.; Gao, Y.; Wei, W. Salient object detection in hyperspectral imagery using multi-scale spectral-spatial gradient. Neurocomputing 2018, 291, 215–225. [Google Scholar] [CrossRef]
  5. Liao, W.; Bellens, R.; Pizurica, A.; Philips, W.; Pi, Y. Classification of Hyperspectral Data Over Urban Areas Using Directional Morphological Profiles and Semi-Supervised Feature Extraction. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 1177–1190. [Google Scholar] [CrossRef]
  6. Gao, L.; Zhao, B.; Jia, X.; Liao, W.; Zhang, B. Optimized Kernel Minimum Noise Fraction Transformation for Hyperspectral Image Classification. Remote Sens. 2017, 9, 548. [Google Scholar] [CrossRef]
  7. Du, L.; You, X.; Li, K.; Meng, L.; Cheng, G.; Xiong, L.; Wang, G. Multi-modal deep learning for landform recognition. ISPRS J. Photogramm. Remote Sens. 2019, 158, 63–75. [Google Scholar] [CrossRef]
  8. Zhang, L.; Shi, Z.; Wu, J. A Hierarchical Oil Tank Detector With Deep Surrounding Features for High-Resolution Optical Satellite Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 4895–4909. [Google Scholar] [CrossRef]
  9. Stankov, K.; He, D.C. Detection of Buildings in Multispectral Very High Spatial Resolution Images Using the Percentage Occupancy Hit-or-Miss Transform. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 4069–4080. [Google Scholar] [CrossRef]
  10. Han, X.; Zhong, Y.; Zhang, L. An Efficient and Robust Integrated Geospatial Object Detection Framework for High Spatial Resolution Remote Sensing Imagery. Remote Sens. 2017, 9, 666. [Google Scholar] [CrossRef]
  11. Pan, X.; Ren, Y.; Sheng, K.; Dong, W.; Yuan, H.; Guo, X.; Ma, C.; Xu, C. Dynamic refinement network for oriented and densely packed object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11207–11216. [Google Scholar]
  12. Li, W.; Chen, Y.; Hu, K.; Zhu, J. Oriented reppoints for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1829–1838. [Google Scholar] [CrossRef]
  13. Han, J.; Ding, J.; Li, J.; Xia, G.S. Align deep features for oriented object detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5602511. [Google Scholar] [CrossRef]
  14. Ming, Q.; Zhou, Z.; Miao, L.; Zhang, H.; Li, L. Dynamic anchor learning for arbitrary-oriented object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 2355–2363. [Google Scholar]
  15. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar] [CrossRef]
  16. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  17. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar] [CrossRef]
  18. Liu, W.; Ma, L.; Wang, J.; Chen, H. Detection of Multiclass Objects in Optical Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2019, 16, 791–795. [Google Scholar] [CrossRef]
  19. Chen, S.; Zhan, R.; Zhang, J. Geospatial Object Detection in Remote Sensing Imagery Based on Multiscale Single-Shot Detector with Activated Semantics. Remote Sens. 2018, 10, 820. [Google Scholar] [CrossRef]
  20. Tang, T.; Zhou, S.; Deng, Z.; Lei, L.; Zou, H. Arbitrary-Oriented Vehicle Detection in Aerial Imagery with Single Convolutional Neural Networks. Remote Sens. 2017, 9, 1170. [Google Scholar] [CrossRef]
  21. Tayara, H.; Chong, K.T. Object Detection in Very High-Resolution Aerial Images Using One-Stage Densely Connected Feature Pyramid Network. Sensors 2018, 18, 3341. [Google Scholar] [CrossRef] [PubMed]
  22. Chen, Z.; Zhang, T.; Ouyang, C. End-to-End Airplane Detection Using Transfer Learning in Remote Sensing Images. Remote Sens. 2018, 10, 139. [Google Scholar] [CrossRef]
  23. Xie, X.; Cheng, G.; Wang, J.; Yao, X.; Han, J. Oriented R-CNN for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 3520–3529. [Google Scholar] [CrossRef]
  24. Cheng, G.; Wang, J.; Li, K.; Xie, X.; Lang, C.; Yao, Y.; Han, J. Anchor-free oriented proposal generator for object detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5625411. [Google Scholar] [CrossRef]
  25. Dai, J.; Li, Y.; He, K.; Sun, J. R-FCN: Object Detection via Region-based Fully Convolutional Networks. In Proceedings of the Advances in Neural Information Processing Systems; Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R., Eds.; Curran Associates, Inc.: New York, NY, USA, 2016; Volume 29. [Google Scholar]
  26. Zhong, Y.; Han, X.; Zhang, L. Multi-class geospatial object detection based on a position-sensitive balancing framework for high spatial resolution remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2018, 138, 281–294. [Google Scholar] [CrossRef]
  27. Chen, C.; Gong, W.; Chen, Y.; Li, W. Object Detection in Remote Sensing Images Based on a Scene-Contextual Feature Pyramid Network. Remote Sens. 2019, 11, 339. [Google Scholar] [CrossRef]
  28. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Region-based convolutional networks for accurate object detection and segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 142–158. [Google Scholar] [CrossRef]
  29. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  31. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  32. Qian, X.; Li, J.; Cao, J.; Wu, Y.; Wang, W. Micro-cracks detection of solar cells surface via combining short-term and long-term deep features. Neural Netw. 2020, 127, 132–140. [Google Scholar] [CrossRef] [PubMed]
  33. Qian, X.; Lin, S.; Cheng, G.; Yao, X.; Ren, H.; Wang, W. Object detection in remote sensing images based on improved bounding box regression and multi-level features fusion. Remote Sens. 2020, 12, 143. [Google Scholar] [CrossRef]
  34. Qian, X.; Huo, Y.; Cheng, G.; Yao, X.; Li, K.; Ren, H.; Wang, W. Incorporating the Completeness and Difficulty of Proposals Into Weakly Supervised Object Detection in Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 1902–1911. [Google Scholar] [CrossRef]
  35. Li, L.; Yao, X.; Wang, X.; Hong, D.; Cheng, G.; Han, J. Robust few-shot aerial image object detection via unbiased proposals filtration. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5617011. [Google Scholar] [CrossRef]
  36. Qian, X.; Wang, C.; Li, C.; Li, Z.; Zeng, L.; Wang, W.; Wu, Q. Multi-Scale Image Splitting Based Feature Enhancement and Instance Difficulty Aware Training for Weakly Supervised Object Detection in Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 7497–7506. [Google Scholar] [CrossRef]
  37. Qian, X.; Huo, Y.; Cheng, G.; Gao, C.; Yao, X.; Wang, W. Mining High-quality Pseudo Instance Soft Labels for Weakly Supervised Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5607615. [Google Scholar] [CrossRef]
  38. Xie, X.; Cheng, G.; Feng, X.; Yao, X.; Qian, X.; Han, J. Attention Erasing and Instance Sampling for Weakly Supervised Object Detection. IEEE Trans. Geosci. Remote Sens. 2023, 62, 5600910. [Google Scholar] [CrossRef]
  39. Qian, X.; Li, C.; Wang, W.; Yao, X.; Cheng, G. Semantic segmentation guided pseudo label mining and instance re-detection for weakly supervised object detection in remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2023, 119, 103301. [Google Scholar] [CrossRef]
  40. Xie, X.; Lang, C.; Miao, S.; Cheng, G.; Li, K.; Han, J. Mutual-assistance learning for object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 15171–15184. [Google Scholar] [CrossRef]
  41. Zheng, X.; Cui, H.; Lu, X. Multiple Source Domain Adaptation for Multiple Object Tracking in Satellite Video. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5626911. [Google Scholar] [CrossRef]
  42. Zheng, X.; Wang, B.; Du, X.; Lu, X. Mutual Attention Inception Network for Remote Sensing Visual Question Answering. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5606514. [Google Scholar] [CrossRef]
  43. Ming, Q.; Miao, L.; Zhou, Z.; Dong, Y. CFC-Net: A critical feature capturing network for arbitrary-oriented object detection in remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5605814. [Google Scholar] [CrossRef]
  44. Qian, X.; Zeng, Y.; Wang, W.; Zhang, Q. Co-saliency detection guided by group weakly supervised learning. IEEE Trans. Multimed. 2022, 25, 1810–1818. [Google Scholar] [CrossRef]
  45. Wang, J.; Li, F.; Bi, H. Gaussian focal loss: Learning distribution polarized angle prediction for rotated object detection in aerial images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4707013. [Google Scholar] [CrossRef]
  46. Li, Z.; Hou, B.; Wu, Z.; Ren, B.; Ren, Z.; Jiao, L. Gaussian synthesis for high-precision location in oriented object detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5619612. [Google Scholar] [CrossRef]
  47. Zhang, C.; Su, J.; Ju, Y.; Lam, K.M.; Wang, Q. Efficient inductive vision transformer for oriented object detection in remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5616320. [Google Scholar] [CrossRef]
  48. Cheng, G.; Li, Q.; Wang, G.; Xie, X.; Min, L.; Han, J. SFRNet: Fine-Grained Oriented Object Recognition via Separate Feature Refinement. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5610510. [Google Scholar] [CrossRef]
  49. Qian, X.; Zhang, N.; Wang, W. Smooth giou loss for oriented object detection in remote sensing images. Remote Sens. 2023, 15, 1259. [Google Scholar] [CrossRef]
  50. Qian, X.; Wu, B.; Cheng, G.; Yao, X.; Wang, W.; Han, J. Building a bridge of bounding box regression between oriented and horizontal object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5605209. [Google Scholar] [CrossRef]
  51. Zheng, X.; Cui, H.; Xu, C.; Lu, X. Dual Teacher: A Semisupervised Cotraining Framework for Cross-Domain Ship Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5613312. [Google Scholar] [CrossRef]
  52. Xie, X.; Cheng, G.; Li, Q.; Miao, S.; Li, K.; Han, J. Fewer is more: Efficient object detection in large aerial images. Science China Inf. Sci. 2024, 67, 112106. [Google Scholar] [CrossRef]
  53. Zheng, X.; Chen, X.; Lu, X. Visible-Infrared Person Re-Identification via Partially Interactive Collaboration. IEEE Trans. Image Process. 2022, 31, 6951–6963. [Google Scholar] [CrossRef] [PubMed]
  54. Yang, X.; Yan, J.; Feng, Z.; He, T. R3det: Refined single-stage detector with feature refinement for rotating object. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 3163–3171. [Google Scholar]
  55. Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning RoI transformer for oriented object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2849–2858. [Google Scholar]
  56. Cheng, G.; Yao, Y.; Li, S.; Li, K.; Xie, X.; Wang, J.; Yao, X.; Han, J. Dual-aligned oriented detector. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5618111. [Google Scholar] [CrossRef]
  57. Yao, Y.; Cheng, G.; Wang, G.; Li, S.; Zhou, P.; Xie, X.; Han, J. On improving bounding box representations for oriented object detection. IEEE Trans. Geosci. Remote Sens. 2022, 61, 5600111. [Google Scholar] [CrossRef]
  58. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  59. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  60. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  61. Neubeck, A.; Van Gool, L. Efficient non-maximum suppression. In Proceedings of the 18th international conference on pattern recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; IEEE: New York, NY, USA, 2006; Volume 3, pp. 850–855. [Google Scholar] [CrossRef]
  62. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hong Kong, China, 20–24 August 2018; pp. 3974–3983. [Google Scholar]
  63. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  64. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open mmlab detection toolbox and benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
  65. Wang, J.; Yang, W.; Li, H.C.; Zhang, H.; Xia, G.S. Learning center probability map for detecting objects in aerial images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4307–4323. [Google Scholar] [CrossRef]
  66. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. Scrdet: Towards more robust detection for small, cluttered and rotated objects. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8232–8241. [Google Scholar] [CrossRef]
  67. Xu, C.; Ding, J.; Wang, J.; Yang, W.; Yu, H.; Yu, L.; Xia, G.S. Dynamic Coarse-to-Fine Learning for Oriented Tiny Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7318–7328. [Google Scholar]
  68. Zhang, C.; Xiong, B.; Li, X.; Kuang, G. TCD: Task-collaborated detector for oriented objects in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4700714. [Google Scholar] [CrossRef]
  69. Qiao, Y.; Miao, L.; Zhou, Z.; Ming, Q. A Novel Object Detector Based on High-quality Rotation Proposal Generation and Adaptive Angle Optimization. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5617715. [Google Scholar] [CrossRef]
  70. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  71. Wu, Y.; Chen, Y.; Yuan, L.; Liu, Z.; Wang, L.; Li, H.; Fu, Y. Rethinking classification and localization for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10186–10195. [Google Scholar] [CrossRef]
Figure 1. Illustration of misclassification problem. (a) Detection result of OOD model without scene classification. (b) Detection result of our method. The blue and yellow rectangles denote tennis courts and a harbor, respectively. The red circle denotes the misclassification object.
Figure 1. Illustration of misclassification problem. (a) Detection result of OOD model without scene classification. (b) Detection result of our method. The blue and yellow rectangles denote tennis courts and a harbor, respectively. The red circle denotes the misclassification object.
Mathematics 12 01343 g001
Figure 2. Framework of the proposed method.
Figure 2. Framework of the proposed method.
Mathematics 12 01343 g002
Figure 3. Parameter analysis of α on DOTA dataset.
Figure 3. Parameter analysis of α on DOTA dataset.
Mathematics 12 01343 g003
Figure 4. Parameter analysis of T on DOTA dataset.
Figure 4. Parameter analysis of T on DOTA dataset.
Mathematics 12 01343 g004
Figure 5. Visualizations of detection results on the DOTA dataset.
Figure 5. Visualizations of detection results on the DOTA dataset.
Mathematics 12 01343 g005
Figure 6. Visualization of detection results on the DIOR-R dataset.
Figure 6. Visualization of detection results on the DIOR-R dataset.
Mathematics 12 01343 g006
Figure 7. Illustration of some failure results of our method. (a) failure results caused by inter-class similarity on the DOTA dataset. (b) failure results caused by shadows on the DIOR-R dataset. The meaning of colors of bounding boxes in (a) and (b) refers to Figure 5 and Figure 6, respectively.
Figure 7. Illustration of some failure results of our method. (a) failure results caused by inter-class similarity on the DOTA dataset. (b) failure results caused by shadows on the DIOR-R dataset. The meaning of colors of bounding boxes in (a) and (b) refers to Figure 5 and Figure 6, respectively.
Mathematics 12 01343 g007
Table 1. Ablation study of SCB and TFL on the DOTA dataset.
Table 1. Ablation study of SCB and TFL on the DOTA dataset.
BaselineSCBTFLmAP
75.24
76.75
76.69
76.95
Table 2. Comparisons with 15 OOD methods on the DOTA dataset under the condition of single-scale training and testing. The AP of each category and the mAP of 15 categories are listed here. The best, sub-optimal and third best results are denoted in red, green and blue colors, respectively (the same below).
Table 2. Comparisons with 15 OOD methods on the DOTA dataset under the condition of single-scale training and testing. The AP of each category and the mAP of 15 categories are listed here. The best, sub-optimal and third best results are denoted in red, green and blue colors, respectively (the same below).
MethodsBackboneEpochPLBDBRGTFSVLVSHTC
Retina-O [60]ResNet501288.6777.6241.8158.1774.5871.6479.1190.29
Faster-RCNN [31]ResNet501288.3473.0644.8659.0973.2571.4977.1190.84
DRN [11]H-10412088.9180.2243.5263.3573.4870.6984.9490.14
CenterMap-Net [65]ResNet501288.8881.2453.1560.6578.6266.5578.1088.83
SCRDet [66]ResNet501289.3480.6552.0968.3660.3272.4190.8587.94
S 2 A-Net [13]ResNet501289.1182.8448.3771.1178.1179.3987.2590.83
DCFL [67]ResNet5012--------
ROI Transformer [55]ResNet501288.6582.6052.5370.8777.9376.6786.8790.71
TCD [68]ResNet501289.2783.7956.9172.1365.7576.7670.6790.88
AOPG [24]ResNet501289.2783.4952.5069.9773.5182.3187.9580.89
DODet [56]ResNet501283.9484.3151.3971.0479.0482.8688.1590.90
Oriented R-CNN [23]ResNet501289.4682.1254.7870.8678.9383.0088.2090.90
Oriented RepPoints [12]ResNet504087.0283.1754.1371.1680.1878.4087.2890.90
OIBBR [57]ResNet501289.5583.6654.0673.9378.9383.0888.2980.89
RPGAOD [69]ResNet501289.3483.5354.3976.5978.0981.4487.6490.83
SCTFL (Ours)ResNet501289.8583.4955.6174.6178.7183.4788.0590.90
MethodsBackboneEpochBCSTSBFRAHASPHCmAP
Retina-O [60]ResNet501282.1874.3254.7560.6062.5769.5760.6468.43
Faster-RCNN [31]ResNet501278.9483.9048.5962.9562.1864.9156.1869.05
DRN [11]H-10412083.8584.1150.1258.4167.6268.6052.5070.70
CenterMap-Net [65]ResNet501277.8083.6149.3666.1972.1072.3658.7071.74
SCRDet [66]ResNet501287.9486.8665.0266.6866.2568.2465.2172.61
S 2 A-Net [13]ResNet501284.9085.6460.3662.6065.2669.1357.9474.12
DCFL [67]ResNet5012-------74.26
ROI Transformer [55]ResNet501283.8382.5153.9567.6174.6768.7561.0374.61
TCD [68]ResNet501286.9583.8160.1671.2176.5471.9570.8975.18
AOPG [24]ResNet501287.6484.7160.0166.1274.1968.3057.8075.24
DODet [56]ResNet501286.8884.9162.6967.6375.4772.2245.5475.49
Oriented R-CNN [23]ResNet501287.5084.6863.9767.6974.9468.8452.2875.87
Oriented RepPoints [12]ResNet504085.9786.2559.9070.4973.5372.2758.9775.97
OIBBR [57]ResNet501286.6084.8062.0365.5574.1670.0958.1676.25
RPGAOD [69]ResNet501258.8985.3365.4464.9673.7370.3159.5076.47
SCTFL (Ours)ResNet501286.8784.3566.3167.1074.5568.2162.2076.95
Table 3. Comparisons with 14 OOD methods on the DIOR-R dataset. The AP of each category and the mAP of 20 categories are listed here.
Table 3. Comparisons with 14 OOD methods on the DIOR-R dataset. The AP of each category and the mAP of 20 categories are listed here.
MethodsBackboneEpoch APLAPOBFBCBRCHDAMETSESAGF
FCOS-O [70]ResNet5012 48.7024.8863.5780.9718.4168.9923.2642.3760.2564.83
RetinaNet-O [60]ResNet5012 61.4928.5273.5781.1723.9872.5419.9472.3958.2069.25
Double-heads [71]ResNet5012 62.1319.5371.5087.0928.0172.1720.3561.1964.5673.37
Faster RCNN-O [31]ResNet5012 62.7926.8071.7280.9134.2072.5718.9566.4565.7566.63
Gliding Vertex [65]ResNet5012 65.3528.8774.9681.3333.8874.3119.5870.7264.7072.30
S 2 A-Net [13]ResNet5012 65.4042.0475.1583.9136.0172.6128.0165.0975.1175.56
ROI Transformer [55]ResNet5012 63.3437.8871.7887.5340.6872.6026.8678.7168.0968.96
OIBBR [57]ResNet5012 63.2241.3971.9788.5541.2372.6328.8278.9069.0070.07
AOPG [24]ResNet5012 62.3937.7971.6287.6340.9072.4731.0865.4277.9973.20
Oriented R-CNN [23]ResNet5012 62.0044.9271.7887.9343.8472.6435.4666.3981.3574.10
TCD [68]ResNet5012 67.9545.3073.4883.5540.3874.7433.6070.5479.3378.03
DODet [56]ResNet5012 63.4043.3572.1181.3243.1272.5933.3278.7770.8474.15
Oriented RepPoints [12]ResNet5012 ----------
DCFL [69]ResNet5012 ----------
SCTFL (Ours)ResNet5012 69.4543.7877.4288.2646.6772.5435.1779.4363.9572.79
MethodsBackboneEpochGTFHAOPSHSTASTOTCTSVEVMmAP
FCOS-O [70]ResNet501250.6631.8440.8073.0966.3256.6177.5538.1030.6955.8751.39
RetinaNet-O [60]ResNet501279.5432.1444.8777.7165.5761.0981.4647.3338.0160.2457.55
Double-heads [71]ResNet501281.9740.6842.4080.3673.1262.3787.0954.9441.3264.8659.45
Faster RCNN-O [31]ResNet501279.2434.9548.7981.1464.3471.2181.4447.3150.4665.2159.54
Gliding Vertex [65]ResNet501278.6837.2249.6480.2269.2661.1381.4944.7647.7165.0460.06
S 2 A-Net [13]ResNet501280.4735.9152.1082.3365.8966.0884.6154.1348.0069.6762.90
ROI Transformer [55]ResNet501282.7447.7155.6181.2178.2370.2681.6154.8643.2765.5263.87
OIBBR [57]ResNet501283.0147.8355.5481.2372.1562.6689.0558.0943.3865.3664.20
AOPG [24]ResNet501281.9442.3254.4581.1772.6971.3181.4960.0452.3869.9964.41
Oriented R-CNN [23]ResNet501280.9543.5258.4281.2568.0165.5288.6259.3143.2766.3164.63
TCD [68]ResNet501278.2248.4656.0085.8965.0768.9981.6157.2248.7263.7965.04
DODet [56]ResNet501275.4748.0059.3185.4174.0471.5681.5255.4751.8666.4065.10
Oriented RepPoints [12]ResNet5012----------66.71
DCFL [69]ResNet5012----------66.80
SCTFL (Ours)ResNet501281.1747.5053.3489.2378.9977.9988.5460.0253.1971.2267.28
Table 4. Comparisons with 7 OOD methods on the DOTA dataset under the condition of multi-scale training and testing. The AP of each category and the mAP of 15 categories are listed here. The bold font denotes the best results (the same below).
Table 4. Comparisons with 7 OOD methods on the DOTA dataset under the condition of multi-scale training and testing. The AP of each category and the mAP of 15 categories are listed here. The bold font denotes the best results (the same below).
MethodsBackboneEpochPLBDBRGTFSVLVSHTC
DRN [11]H-10412089.4583.1648.9862.2470.6374.2583.9990.73
S 2 A-Net [13]ResNet501288.8983.6057.7481.9579.9483.1989.1190.78
TCD [68]ResNet501271.7780.5658.1889.7888.3177.8483.8868.04
AOPG [24]ResNet501289.8885.5760.9081.5178.7085.2988.8590.89
DODet [56]ResNet501289.9685.5258.0181.2278.7185.4688.5990.89
Oriented R-CNN [23]ResNet501289.8485.4361.0979.8279.7185.3588.8290.88
OIBBR [57]ResNet501290.1485.3160.9879.9280.2185.0488.8090.87
SCTFL (Ours)ResNet501289.7984.6760.9779.3979.3185.4688.3690.88
MethodsBackboneEpochBCSTSBFRAHASPHCmAP
DRN [11]H-10412084.6085.3555.7660.7971.5668.8263.9272.95
S 2 A-Net [13]ResNet501284.8787.8170.3068.2578.3077.0169.5879.41
TCD [68]ResNet501286.2978.9290.8478.9772.5088.0186.8580.05
AOPG [24]ResNet501287.6087.6571.6668.6982.3177.3273.1080.66
DODet [56]ResNet501287.1287.8070.5071.5482.0677.4345.5474.47
Oriented R-CNN [23]ResNet501286.6887.7372.2170.8082.4278.1874.1180.87
OIBBR [57]ResNet501286.4588.0470.8871.7282.9980.5573.1981.00
SCTFL (Ours)ResNet501287.2887.3971.2572.0483.3581.9874.2081.09
Table 5. Comparisons with 14 OOD methods in the densely populated scenes of the DOTA dataset. The AP of each scene and the mAP of 4 scenes are listed here.
Table 5. Comparisons with 14 OOD methods in the densely populated scenes of the DOTA dataset. The AP of each scene and the mAP of 4 scenes are listed here.
MethodsSVLVSHTCmAP
SCRDet [66]68.3660.3272.4190.8572.98
TCD [68]65.7576.7670.6790.8876.01
CenterMap-Net [65]78.6266.5578.1088.8378.02
Faster RCNN [31]73.2571.4977.1190.8478.17
Retina-O [60]74.5871.6479.1190.2978.90
DRN [11]73.4870.6984.9490.1479.81
ROI Transformer [55]77.9376.6786.8790.7183.04
S 2 A-Net [13]78.1178.3987.2590.8383.64
AOPG [24]73.5182.3187.9590.8983.66
Oriented RepPoints [12]80.1878.4087.2890.9084.19
RPGAOD [69]78.0981.4487.6490.8384.50
DODet [56]79.0482.8688.1590.9085.23
Oriented R-CNN [23]78.9383.0088.2090.9085.25
OIBBR [57]78.9383.0888.2990.8985.29
SCTFL (Ours)78.6183.6688.2790.9085.36
Table 6. Comparisons with 12 OOD methods in the densely populated scenes of the DIOR-R dataset. The AP of each scene and the mAP of 4 scenes are listed here.
Table 6. Comparisons with 12 OOD methods in the densely populated scenes of the DIOR-R dataset. The AP of each scene and the mAP of 4 scenes are listed here.
MethodsAPLSHTCVEmAP
FCOS-O [70]48.7073.0977.5530.6957.50
RetinaNet-O [60]61.4977.7181.4638.0164.66
ROI Transformer [55]63.3481.2181.6143.2767.35
Double-heads [71]62.1380.3687.0941.3267.72
Gliding Vertex [65]65.3580.2281.4947.7168.69
Oriented R-CNN [23]62.0081.2588.6243.2768.78
Faster RCNN-O [31]62.7981.1481.4450.4668.95
OIBBR [57]63.2281.2389.0543.3869.22
AOPG [24]62.3981.1781.4952.3869.35
S 2 A-Net [13]65.4082.3384.6148.0070.08
DODet [56]63.4085.4181.5251.8670.54
TCD [68]67.9585.8981.6148.7271.04
SCTFL (Ours)69.4589.2388.5453.1975.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qian, X.; Gao, S.; Deng, W.; Wang, W. Improving Oriented Object Detection by Scene Classification and Task-Aligned Focal Loss. Mathematics 2024, 12, 1343. https://doi.org/10.3390/math12091343

AMA Style

Qian X, Gao S, Deng W, Wang W. Improving Oriented Object Detection by Scene Classification and Task-Aligned Focal Loss. Mathematics. 2024; 12(9):1343. https://doi.org/10.3390/math12091343

Chicago/Turabian Style

Qian, Xiaoliang, Shaoguan Gao, Wei Deng, and Wei Wang. 2024. "Improving Oriented Object Detection by Scene Classification and Task-Aligned Focal Loss" Mathematics 12, no. 9: 1343. https://doi.org/10.3390/math12091343

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop