Laser & Optoelectronics Progress, Volume. 58, Issue 2, 0228002(2021)

Classification of Individual Tree Species in High-Resolution Remote Sensing Imagery Based on Convolution Neural Network

Guang Ouyang1,2, Linhai Jing1、*, Shijie Yan1, Hui Li1, Yunwei Tang1, and Bingxiang Tan3
Author Affiliations
  • 1Key Laboratory of Digital Earth, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
  • 2School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Science, Beijing 100049, China
  • 3Institute of Forest Resource Information Techniques CAF, Beijing 100091, China
  • show less
    Figures & Tables(14)
    Location of the research area. (a) Huangshan City, Anhui Province; (b) true color schematic of WorldView3, the box indicates the location of Huangshan Mountain
    Construction steps of sample set of remote sensing imagery of individual tree species. (a) Remote sensing imagery of research area; (b) distribution diagram of tree species; (c) delineation diagram of tree crown; (d) labeling diagram of tree crown category; (e) remote sensing imagery of individual tree species; (f) sample set of remote sensing imagery of individual tree species
    Classification labeling result of sample set of remote sensing imagery of individual tree species
    Histogram of training accuracy, validation accuracy, and network layers when CNN model converges
    Classification diagram of tree species of Huangshan Mountain
    • Table 1. Results of sample set division before and after data augmentation

      View table

      Table 1. Results of sample set division before and after data augmentation

      Tree speciesTraining sample setValidation sample setTest sample set
      BeforeAfterBeforeAfter
      Ph.h663962313823
      E.a26615968953489
      C.l834982816828
      Pi.h119971944012406401
      D.a3692214124744124
      Total1983118986653990665
    • Table 2. LeNet5_relu model parameter

      View table

      Table 2. LeNet5_relu model parameter

      LayerOutput sizeParameter
      Input32×32×8-
      Convolutional C128×28×6Kernel 5×5, filter 6, stride 1, ReLU
      Pooling S114×14×6Average_pooling 2×2, stride 2
      Convolutional C210×10×16Kernel 5×5, filter 16, stride 1, ReLU
      Pooling S25×5×16Average_pooling 2×2, stride 2
      Convolutional C31×1×120Kernel 5×5, filter 120, stride 1, ReLU
      Fully-connected F184Node 84, FC, ReLU
      Classification5Node 5, FC, Softmax
    • Table 3. AlexNet_mini model parameter

      View table

      Table 3. AlexNet_mini model parameter

      LayerOutput sizeParameter
      Input32×32×8-
      Convolutional C132×32×12Kernel 7×7, filter 12, stride 1, ReLU
      Pooling S115×15×12Average_pooling 3×3, stride 2
      Convolutional C215×15×36Kernel 5×5, filter 36, stride 1, ReLU
      Pooling S27×7×36Average_pooling 3×3, stride 2
      Convolutional C37×7×54Kernel 3×3, filter 54, stride 1, ReLU
      Convolutional C47×7×54Kernel 3×3, filter 54, stride 1, ReLU
      Convolutional C53×3×36Kernel 3×3, filter 36, stride 1, ReLU
      Pooling S33×3×36Average_pooling 3×3, stride 2
      Fully-connected F1320Node 320, FC, ReLU
      Fully-connected F2100Node 100, FC, ReLU
      Classification5Node 5, FC, Softmax
    • Table 4. GoogLeNet_mini56 model parameter

      View table

      Table 4. GoogLeNet_mini56 model parameter

      LayerOutput sizeParameter
      Input32×32×8--
      Convolutional C132×32×12Kernel 7×7, filter 12, stride 1
      Inception V1 block (1a)32×32×32--
      Inception V1 block (1b)32×32×60--
      Pooling S115×15×60Max_pooling 3×3, stride 2
      Inception V1 block (2a)15×15×64--
      Inception V1 block (2b)15×15×64--
      Inception V1 block (2c)15×15×64--
      Inception V1 block (2d)15×15×66--
      Inception V1 block (2e)15×15×104--
      Pooling S27×7×104Max_pooling 3×3, stride 2
      Inception V1 block (3a)7×7×104--
      Inception V1 block (3b)7×7×128--
      Pooling S31×1×128Average_pooling 7×7, stride 1
      Classification5Node 5, FC, Softmax
    • Table 5. Inception V1 block parameter

      View table

      Table 5. Inception V1 block parameter

      LayerInception V1 block
      Convolutionalkernel 1×1Bottleneckkernel 1×1Convolutionalkernel 3×3Bottleneckkernel 1×1Convolutionalkernel 5×5Pooling3×3Bottleneckkernel 1×1
      1aFilter 8Filter 12Filter 16Filter 2Filter 4Stride 1Filter 4
      1bFilter 16Filter 16Filter 24Filter 4Filter 12Stride 1Filter 8
      2aFilter 24Filter 12Filter 26Filter 2Filter 6Stride 1Filter 8
      2bFilter 20Filter 14Filter 28Filter 3Filter 8Stride 1Filter 8
      2cFilter 16Filter 16Filter 32Filter 3Filter 8Stride 1Filter 8
      2dFilter 14Filter 18Filter 36Filter 4Filter 8Stride 1Filter 8
      2eFilter 32Filter 20Filter 40Filter 4Filter 16Stride 1Filter 16
      3aFilter 32Filter 20Filter 40Filter 4Filter 16Stride 1Filter 16
      3bFilter 48Filter 24Filter 48Filter 6Filter 16Stride 1Filter 16
    • Table 6. ResNet_mini56 model parameter

      View table

      Table 6. ResNet_mini56 model parameter

      LayerOutput sizeParameterResidual blockfilter
      Bottleneckkernel 1×1Convolutionalkernel 3×3Bottleneckkernel 1×1
      Input32×32×8----
      Convolutional C116×16×12Kernel 7×7, filter 12, stride 2--
      Residual block(1)Pooling S18×8×12Max_pooling 3×3, stride 2--
      Bottleneck8×8×12× 312----
      Convolutional8×8×12--12--
      Bottleneck8×8×48----48
      Residual block(2)Bottleneck4×4×24× 424----
      Convolutional4×4×24--24--
      Bottleneck4×4×96----96
      Residual block(3)Bottleneck2×2×48×848----
      Convolutional2×2×48--48--
      Bottleneck2×2×192----192
      Residual block(4)Bottleneck1×1×96× 396----
      Convolutional1×1×96--96--
      Bottleneck1×1×384----384
      Classification384Global_average_pooling--
      5Node 5, FC,Softmax
    • Table 7. DenseNet_BC_mini56 model parameter

      View table

      Table 7. DenseNet_BC_mini56 model parameter

      LayerOutput sizeParameterDense block filter
      Bottleneckkernel 1×1Convolutionalkernel 1×1
      Input32×32×8----
      Convolutional C132×32×12Kernel 3×3, filter 12, stride 2--
      Dense block(1)Bottleneck32×32×42×524--
      Convolutional--6
      Compression(1)32×32×21Kernel 1×1--
      16×16×21Average_pooling 2×2, stride 2
      Dense block(2)Bottleneck16×16×51× 524--
      Convolutional--6
      Compression(2)16×16×36Kernel 1×1--
      8×8×36Average_pooling 2×2, stride 2
      Dense block(3)Bottleneck8×8×66×524--
      Convolutional--6
      Compression(3)8×8×51Kernel 1×1--
      4×4×51Average_pooling 2×2, stride 2
      Dense block(4)Bottleneck4×4×81× 524--
      Convolutional--6
      Compression(4)4×4×66Kernel 1×1--
      2×2×66Average_pooling 2×2, stride 2
      Dense block(5)Bottleneck2×2×99×524--
      Convolutional--6
      Classification99Global_average_pooling--
      5Node 5, FC, Softmax
    • Table 8. CNN model parameter

      View table

      Table 8. CNN model parameter

      Model nameTotal parameterTrainable parameterNon-trainable parameterNetwork layer
      LeNet5_relu623316233105
      AlexNet_mini21353721353708
      GoogLeNet_mini5697251972272456
      ResNet_mini56934025924401962456
      DenseNet_BC_mini568297978839414056
    • Table 9. Classification accuracy evaluation index of CNN model

      View table

      Table 9. Classification accuracy evaluation index of CNN model

      Model nameEvaluation indexTree species
      Ph.hE.aC.lPi.hD.a
      LeNet5_reluProducer accuracy /%86.9667.4275.0098.0087.90
      User accuracy /%90.9176.9287.5093.5790.08
      Overall accuracy /%90.68
      Kappa coefficient0.84
      AlexNet_miniProducer accuracy /%86.9671.9164.2997.7692.74
      User accuracy /%90.9179.0178.2694.0094.26
      Overall accuracy /%91.58
      Kappa coefficient0.85
      GoogLeNet_mini56Producer accuracy /%95.6574.1675.0097.7697.58
      User accuracy /%100.0084.6295.4594.9293.08
      Overall accuracy /%93.53
      Kappa coefficient0.89
      ResNet_mini56Producer accuracy /%95.6576.4078.5797.5192.74
      User accuracy /%100.0080.95100.0094.9092.00
      Overall accuracy /%92.93
      Kappa coefficient0.88
      DenseNet_BC_mini56Producer accuracy /%95.6575.2885.7198.2595.97
      User accuracy /%100.0087.01100.0094.7194.44
      Overall accuracy /%94.14
      Kappa coefficient0.90
    Tools

    Get Citation

    Copy Citation Text

    Guang Ouyang, Linhai Jing, Shijie Yan, Hui Li, Yunwei Tang, Bingxiang Tan. Classification of Individual Tree Species in High-Resolution Remote Sensing Imagery Based on Convolution Neural Network[J]. Laser & Optoelectronics Progress, 2021, 58(2): 0228002

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Remote Sensing and Sensors

    Received: Jun. 12, 2020

    Accepted: Jul. 3, 2020

    Published Online: Jan. 11, 2021

    The Author Email: Jing Linhai (jinglh@radi.ac.cn)

    DOI:10.3788/LOP202158.0228002

    Topics