Figure 1.
Framework of the proposed multi-screening unlabeled sample semi-supervised learning method.
Figure 1.
Framework of the proposed multi-screening unlabeled sample semi-supervised learning method.
Figure 2.
Framework of unlabeled sample labeling with early training multi-models.
Figure 2.
Framework of unlabeled sample labeling with early training multi-models.
Figure 3.
Framework of high-probability sample selection.
Figure 3.
Framework of high-probability sample selection.
Figure 4.
Framework of the retraining and small loss selection.
Figure 4.
Framework of the retraining and small loss selection.
Figure 5.
Some examples from the NWPU-RESISC45 dataset. (0) Airplane, (1) airport, (2) baseball diamond, (3) basketball court, (4) beach, (5) bridge, (6) chaparral, (7) church, (8) circular farmland, (9) cloud, (10) commercial area, (11) dense residential, (12) desert, (13) forest, (14) freeway, (15) golf course, (16) ground track field, (17) harbor, (18) industrial area, (19) intersection, (20) island, (21) lake, (22) meadow, (23) medium residential, (24) mobile home park, (25) mountain, (26) overpass, (27) palace, (28) parking lot, (29) railway, (30) railway station, (31) rectangular farmland, (32) river, (33) roundabout, (34) runway, (35) sea ice, (36) ship, (37) snow berg, (38) sparse residential, (39) stadium, (40) storage tank, (41) tennis court, (42) terrace, (43) thermal power station, (44) wetland.
Figure 5.
Some examples from the NWPU-RESISC45 dataset. (0) Airplane, (1) airport, (2) baseball diamond, (3) basketball court, (4) beach, (5) bridge, (6) chaparral, (7) church, (8) circular farmland, (9) cloud, (10) commercial area, (11) dense residential, (12) desert, (13) forest, (14) freeway, (15) golf course, (16) ground track field, (17) harbor, (18) industrial area, (19) intersection, (20) island, (21) lake, (22) meadow, (23) medium residential, (24) mobile home park, (25) mountain, (26) overpass, (27) palace, (28) parking lot, (29) railway, (30) railway station, (31) rectangular farmland, (32) river, (33) roundabout, (34) runway, (35) sea ice, (36) ship, (37) snow berg, (38) sparse residential, (39) stadium, (40) storage tank, (41) tennis court, (42) terrace, (43) thermal power station, (44) wetland.
Figure 6.
Confusion matrix obtained by our proposed method on NWPU-RESISC45 testing set.
Figure 6.
Confusion matrix obtained by our proposed method on NWPU-RESISC45 testing set.
Figure 7.
Confusion matrix obtained by MixMatch on NWPU-RESISC45 testing set.
Figure 7.
Confusion matrix obtained by MixMatch on NWPU-RESISC45 testing set.
Figure 8.
Some examples from the AID dataset. (0) Airport, (1) bare land, (2) baseball field, (3) beach, (4) bridge, (5) center, (6) church, (7) commercial, (8) dense residential, (9) desert, (10) farmland, (11) forest, (12) industrial, (13) meadow, (14) medium residential, (15) mountain, (16) park, (17) parking, (18) playground, (19) pond, (20) port, (21) railway station, (22) resort, (23) river, (24) school, (25) sparse residential, (26) square, (27) stadium, (28) storage tank, (29) viaduct.
Figure 8.
Some examples from the AID dataset. (0) Airport, (1) bare land, (2) baseball field, (3) beach, (4) bridge, (5) center, (6) church, (7) commercial, (8) dense residential, (9) desert, (10) farmland, (11) forest, (12) industrial, (13) meadow, (14) medium residential, (15) mountain, (16) park, (17) parking, (18) playground, (19) pond, (20) port, (21) railway station, (22) resort, (23) river, (24) school, (25) sparse residential, (26) square, (27) stadium, (28) storage tank, (29) viaduct.
Figure 9.
Confusion matrix obtained by our proposed method on AID testing set.
Figure 9.
Confusion matrix obtained by our proposed method on AID testing set.
Figure 10.
Precision comparison for each class of our proposed method and the MixMatch method on AID dataset, where the Y-axis denotes per class classification accuracy improvement of our proposed method relative to MixMatch, and the X-axis denotes the class index of each category.
Figure 10.
Precision comparison for each class of our proposed method and the MixMatch method on AID dataset, where the Y-axis denotes per class classification accuracy improvement of our proposed method relative to MixMatch, and the X-axis denotes the class index of each category.
Figure 11.
Sample images of some categories in NWPU-TESISC45 and AID datasets.
Figure 11.
Sample images of some categories in NWPU-TESISC45 and AID datasets.
Table 1.
Recognition accuracy (%) of label propagation, EL + LR, Mean-teacher, ICT, MixMatch, and our proposed method with differently labeled data for each category on NWPU-RESISC45 dataset.
Table 1.
Recognition accuracy (%) of label propagation, EL + LR, Mean-teacher, ICT, MixMatch, and our proposed method with differently labeled data for each category on NWPU-RESISC45 dataset.
Method | NWPU-RESISC45 |
---|
Num of Labeled Data for Each Category |
---|
1 | 2 | 3 | 5 |
---|
Label Propagation [15] | 20.56 | 38.66 | 53.42 | 69.37 |
EL + LR [15] | 25.36 | 40.62 | 68.57 | 70.82 |
Mean-Teacher [16] | 20.01 | 34.33 | 40.31 | 51.34 |
ICT [18] | 40.53 | 62.47 | 70.96 | 77.09 |
MixMatch [19] | 46.46 | 66.66 | 74.26 | 83.76 |
RS [24] | 48.16 | 72.34 | 81.03 | 86.61 |
Ours | 52.91 | 78.26 | 85.17 | 90.34 |
Table 2.
Different semantic categories and corresponding number of images in each type of AID dataset.
Table 2.
Different semantic categories and corresponding number of images in each type of AID dataset.
Datasets | Types | Num | Types | Num | Types | Num |
---|
AID | Airport | 360 | Bare land | 310 | Baseball field | 220 |
Beach | 400 | Bridge | 360 | Center | 260 |
Church | 240 | Commercial | 350 | Dense | 410 |
Desert | 300 | Farmland | 370 | Forest | 250 |
Industrial | 390 | Meadow | 280 | Medium residential | 290 |
Mountain | 340 | Park | 350 | Parking | 390 |
Play ground | 370 | Pond | 420 | Port | 380 |
Railway station | 260 | Resort | 290 | River | 410 |
School | 300 | Sparse residential | 300 | Square | 330 |
Stadium | 290 | Storage tanks | 360 | Viaduct | 420 |
Table 3.
Recognition accuracy (%) of label Propagation, EL + LR, Mean-teacher, ICT, MixMatch, RS, and our proposed method with different labeled data for each category on AID dataset.
Table 3.
Recognition accuracy (%) of label Propagation, EL + LR, Mean-teacher, ICT, MixMatch, RS, and our proposed method with different labeled data for each category on AID dataset.
Method | AID |
---|
Num of Labeled Data for Each Category |
---|
1 | 2 | 3 | 5 |
---|
Label Propagation [15] | 31.24 | 40.21 | 65.71 | 73.42 |
EL + LR [15] | 29.27 | 45.32 | 73.63 | 79.41 |
Mean-Teacher [16] | 19.38 | 31.31 | 40.02 | 51.66 |
ICT [18] | 44.70 | 69.98 | 80.33 | 85.24 |
MixMatch [19] | 48.66 | 74.72 | 85.80 | 91.63 |
RS [24] | 54.13 | 78.92 | 89.13 | 91.82 |
Ours | 61.60 | 83.15 | 91.06 | 94.40 |
Table 4.
Recognition accuracy (%) of the aerial scene classification using different numbers of training epochs for early training of three models.
Table 4.
Recognition accuracy (%) of the aerial scene classification using different numbers of training epochs for early training of three models.
| 10 | 20 | 30 | 40 |
---|
Ours | 83.15 | 81.02 | 79.75 | 78.65 |
Table 5.
Recognition accuracy (%) of the aerial scene classification using different selected numbers of samples for each category.
Table 5.
Recognition accuracy (%) of the aerial scene classification using different selected numbers of samples for each category.
| 0 | 1 | 2 | 3 | 4 | 5 | 6 |
---|
Ours | 75.87 | 77.57 | 78.36 | 80.45 | 83.15 | 81.25 | 68.70 |
Table 6.
Recognition accuracy (%) of the aerial scene classification using different numbers of early training models M.
Table 6.
Recognition accuracy (%) of the aerial scene classification using different numbers of early training models M.
M | 1 | 2 | 3 | 4 | 5 |
---|
Ours | 76.74 | 80.12 | 83.15 | 83.20 | 80.75 |
Table 7.
Time analysis of the proposed method, Mean-teacher, ICT, MixMatch, and RS.
Table 7.
Time analysis of the proposed method, Mean-teacher, ICT, MixMatch, and RS.
| Training Time per Epoch | Testing Time per Image |
---|
Mean-teacher | 98.04 s | 4.60 ms |
ICT | 102.55 s | 4.72 ms |
MixMatch | 112.60 s | 4.58 ms |
RS | 123.46 s | 4.76 ms |
Ours | 131.87 s | 4.83 ms |