
For trees, leaves are often used for identification, but the shape of leaves changes greatly, bark will be another identifying feature. However, it is difficult to recognize by a single organ when there are intra class differences and inter class similarities between leaves or bark. So we fuse features of leaf and bark. Firstly, we collected 17 species of leaves and bark of trees through field shooting and web crawling. Then propose a method of combining convolution neural network (CNN) with cascade fusion, additive fusion algorithm, bilinear fusion and score level fusion. Finally, the features extracted from the leaves and bark are fused in the ReLu layer and Fully connected layer. The method was compared with single organ recognition, Support Vector Machines (SVM), and existing fusion methods, results show that the two organ fusion method proposed are better than the other recognition methods, and recognition accuracy is 87.86%. For similar trees, when it is impossible to accurately determine its species by a single organ, the fusion of two organs can effectively improve this situation.
Citation: Yafeng Zhao, Xuan Gao, Junfeng Hu, Zhen Chen. Tree species identification based on the fusion of bark and leaves[J]. Mathematical Biosciences and Engineering, 2020, 17(4): 4018-4033. doi: 10.3934/mbe.2020222
[1] | Cong Lin, Yiquan Huang, Wenling Wang, Siling Feng, Mengxing Huang . Lesion detection of chest X-Ray based on scalable attention residual CNN. Mathematical Biosciences and Engineering, 2023, 20(2): 1730-1749. doi: 10.3934/mbe.2023079 |
[2] | Xiaochen Liu, Weidong He, Yinghui Zhang, Shixuan Yao, Ze Cui . Effect of dual-convolutional neural network model fusion for Aluminum profile surface defects classification and recognition. Mathematical Biosciences and Engineering, 2022, 19(1): 997-1025. doi: 10.3934/mbe.2022046 |
[3] | Andrew L. Nevai, Richard R. Vance . The role of leaf height in plant competition for sunlight: analysis of a canopy partitioning model. Mathematical Biosciences and Engineering, 2008, 5(1): 101-124. doi: 10.3934/mbe.2008.5.101 |
[4] | Youcef Mammeri, Damien Sellier . A surface model of nonlinear, non-steady-state phloem transport. Mathematical Biosciences and Engineering, 2017, 14(4): 1055-1069. doi: 10.3934/mbe.2017055 |
[5] | Sidra Abid Syed, Munaf Rashid, Samreen Hussain, Anoshia Imtiaz, Hamnah Abid, Hira Zahid . Inter classifier comparison to detect voice pathologies. Mathematical Biosciences and Engineering, 2021, 18(3): 2258-2273. doi: 10.3934/mbe.2021114 |
[6] | Jian Zhang, Yan Zhang, Cong Wang, Huilong Yu, Cui Qin . Binocular stereo matching algorithm based on MST cost aggregation. Mathematical Biosciences and Engineering, 2021, 18(4): 3215-3226. doi: 10.3934/mbe.2021160 |
[7] | Keying Du, Liuyang Fang, Jie Chen, Dongdong Chen, Hua Lai . CTFusion: CNN-transformer-based self-supervised learning for infrared and visible image fusion. Mathematical Biosciences and Engineering, 2024, 21(7): 6710-6730. doi: 10.3934/mbe.2024294 |
[8] | Liangyu Yang, Tianyu Shi, Jidong Lv, Yan Liu, Yakang Dai, Ling Zou . A multi-feature fusion decoding study for unilateral upper-limb fine motor imagery. Mathematical Biosciences and Engineering, 2023, 20(2): 2482-2500. doi: 10.3934/mbe.2023116 |
[9] | Ekaterina Y. Kasap, Оlga K. Parfenova, Roman V. Kurkin, Dmitry V. Grishin . Bioinformatic analysis of the coding region of the melatonin receptor 1b gene as a reliable DNA marker to resolve interspecific mammal phylogenetic relationships. Mathematical Biosciences and Engineering, 2023, 20(3): 5430-5447. doi: 10.3934/mbe.2023251 |
[10] | Jie Bai, Heru Xue, Xinhua Jiang, Yanqing Zhou . Recognition of bovine milk somatic cells based on multi-feature extraction and a GBDT-AdaBoost fusion model. Mathematical Biosciences and Engineering, 2022, 19(6): 5850-5866. doi: 10.3934/mbe.2022274 |
For trees, leaves are often used for identification, but the shape of leaves changes greatly, bark will be another identifying feature. However, it is difficult to recognize by a single organ when there are intra class differences and inter class similarities between leaves or bark. So we fuse features of leaf and bark. Firstly, we collected 17 species of leaves and bark of trees through field shooting and web crawling. Then propose a method of combining convolution neural network (CNN) with cascade fusion, additive fusion algorithm, bilinear fusion and score level fusion. Finally, the features extracted from the leaves and bark are fused in the ReLu layer and Fully connected layer. The method was compared with single organ recognition, Support Vector Machines (SVM), and existing fusion methods, results show that the two organ fusion method proposed are better than the other recognition methods, and recognition accuracy is 87.86%. For similar trees, when it is impossible to accurately determine its species by a single organ, the fusion of two organs can effectively improve this situation.
As one of the most important natural forest distribution areas in China, Northeast China is particularly important for the protection of trees. It is of great significance for plant diversity protection and botany research to realize automatic identification and classification of trees. Tree identification mainly uses leaves, flowers and bark [1,2,3]. There are also many difficulties in the study of tree species identification, such as [4]: (1) There is a lot of characteristic information to distinguish tree species: plant shape, texture, color, etc. (2) There are a lot of noise interference in the plant image under complex background. (3) The same tree will be different in different growth stages and seasons. (4) Tree species image database is incomplete. In order to solve this problem, researchers have done a lot of research on tree species identification and made some progress.
For tree species identification, many methods are proposed, most of which use leaves as recognition organs, and some scholars use bark as recognition organs. The main recognition methods are divided into two categories, which are recognition based on traditional algorithm and recognition based on deep learning.
Traditional algorithms are used to identify and classify trees, such as using AdaBoost, KNN, SVM, etc. [5,6,7]. In [8], using the combination features both texture features and shape feature and proposing a pre-training method based on the PID to improve the DBNs. In [9], the authors proposed a method of blade recognition based on the combination of clonal selection algorithm and support vector machine. In [10], the authors presented a leaf recognition system using orthogonal moments as shape descriptors and Histogram of oriented gradients (HOG) and Gabor features as texture descriptors. In [11], the authors proposed a novel statistical radial binary pattern (SRBP) descriptor to encode the between-scale texture information within large neighborhood areas using the statistical description of the grey scale intensity distribution. In [12], the authors proposed a hierarchical architectural design and another Feature based Shape Selection Template (FSST). Moreover, a novel approach is proposed by using the combination of fuzzy-color and edge-texture histogram in order to recognize fragmented leaf images in 2019 [13]. Traditional recognition algorithms need to select features manually, which has limitations and subjectivity. In addition, the recognition accuracy of traditional algorithms is sometimes lower than that of deep learning algorithms.
Due to boosting in data availability, and accompanying by substantial progress in machine learning algorithms, notably CNN, pushed these approaches to a stage where they are better, faster, cheaper and have the potential to significantly contribute to biodiversity and conservation research [14]. For example, MATHIEU et al. [15] show that there is no dataset with bark images, so a dataset called Baknet1.0 is provided, which contains 23 kinds of trees. And deep learning was used to demonstrate the feasibility of species identification through bark images. In [16], using CNN to realize tree species recognition through leaf images, and additional preprocessing steps are employed to increase the robustness of the identification results, this approach is evaluated based on the Leafsnap database and achieves satisfying performance. In [17], improving the structure of CNN, ELU excitation function with Maxout was used instead of the ReLU function to solve the model offset and zero gradient problem, and the model is verified on 5 kinds of 10, 000 bark images. Zhou et al. [18] provided an effective approach to automatically identify tree species using CNN. The work identifies tree species by analyzing tree leaves, which have multi-dimensional features such as color, shape, and leaf vein signatures. In [19], the authors proposed a plant recognition algorithm based on the optimized P-AlexNet model, the model training uses an image dataset with 206 plants, composed of Oxford102 and Ecust104 dataset, and the validation accuracy of the model is 86.7%. In [20], a deep learning framework is developed to enable path-based tree classifier training for supporting large-scale plant species recognition, where a deep neural network and a tree classifier are jointly trained in an end-to-end fashion. Compared with the traditional algorithm, Deep learning improves the recognition of accuracy and speed. To sum up, many scholars only use leaves or bark for identification. However, the appearance of a certain plant at different growth periods is different, which is called intra class difference, and the appearance similarity of different plants is called inter class similarity, as shown in Figure 1. And even for experienced botanists, sometimes it is impossible to provide a definite identification based on a single image.
At present, some scholars fuse multiple organs to identify plant species. Guo et al. [21] used the framework of CNN, and made the final plant decision based on multiple organs using Linear Weighted Classification and SVM on the dataset of 100 species. Sarah et al. [22] proposed a fusion method based on SVM including belief function and fusion of leaves and bark, compared them on a public database of 72 species of trees and shrubs. Rzanny et al. [23] collected a completely balanced dataset comprising images of leaf and flower for each of 101 species with an emphasis on groups of conspecific and visually similar species including twelve Poaceae species, using this dataset to train CNN and determining the prediction accuracy for each single perspective and their combinations via score level fusion.
Northeast China is chosen as the research area. Because trees of this area have a half year long leaf fall period, so we choose to collect leaf images in spring and summer. The collection of bark image follows the growth cycle of trees, in order to show that there is intra class difference between trees. The experiment is mainly conducted in two aspects. On the one hand, we train two classifiers for all the leaf images and bark images through CNN. On the other hand, we train a classifier for images of leaf and bark of each tree species through fusion. In the subsequent analysis, the results of the second method will be compared with the first method, and also with baseline approach, that is, the established plant identification system (such as Pl@ntNet [24], iNaturalist [25]).We combine fusion algorithm with CNN and improved, so that leaves and bark can be better fused. It can promote the research of forestry development, achieve rapid and efficient identification, and more widely applied to the public.
The training model needs to have qualified training images. Today, people can capture, share images and complete recognition through smartphone APP. For example, Pl@ntNet and iNaturalist have a lot of image data. However, there is currently no image that contains both leaves and bark from the same tree, such images also inhibit a wide range of quality. A widely known example is the PlantCLEF dataset [26], which is used as benchmark for various computer vision tasks [27]. Yet, it is not clear how the results achieved on such a dataset are affected by data imbalance towards image number per species and organs, poor image quality and misidentified species [28]. Therefore, we choose self-photographing and web crawling to accomplish the collection of datasets.
Harbin, Heilongjiang Province, Northeast China, is selected as the research area, with abundant tree species. We selected representative trees and classified them according to type. The 10 kinds of trees were divided into 7 families. In order to ensure the diversity of data and better fit the actual situation. Firstly, we use different mobile phones to take photos, including Apple, Meizu and Huawei. Then we choose different scenes to take photos, including the campus and forest farm of Northeast Forestry University in China. Thirdly, we choose different angles and distances, including parallel and 45 degree elevation taking and between 20 cm and 40 cm from trees. Table 1 shows the species and number of images. The number of leaves and bark of each tree is 400 images respectively. We collect images in three ways, the number of images mainly includes: before expansion (original dataset): 2000 bark images and 2000 leaf images, after expansion (flip and translation): 14000 bark images and 14000 leaf images, web crawling: 838 bark images and 1157 leaf images. Figure 2 shows the images of bark and leaf.
Name | Species | Families | Bark | Leaf | ||||
Before | After | Web crawling | Before | After | Web crawling | |||
1 | Betula platyphylla suk (BP) | Betulaceae | 400 | 1400 | 400 | 1400 | ||
2 | Alnus glutinosa (L.) Gaertn (AG) | 221 | 234 | |||||
3 | Juglans mandshurica maxim (JM) | Juglandaceae | 400 | 1400 | 400 | 1400 | ||
4 | Carya glabra (Mill.) Sweet (CG) | 35 | 76 | |||||
5 | Juglans regia L (JR) | 283 | 306 | |||||
6 | Quercus cerris L (QC) | Fagaceae | 61 | 179 | ||||
7 | Quercus mongolica fisch (QM) | 400 | 1400 | 400 | 1400 | |||
8 | Raxinus mandshurica rupr (FM) | Oleaceae | 400 | 1400 | 400 | 1400 | ||
9 | Fraxinus angustifolia Vahl (FA) | 140 | 195 | |||||
10 | Phellodendron amurense Rupr (PA) | Rutaceae | 400 | 1400 | 400 | 1400 | ||
11 | Tilia amurensis Rupr (TA) | Tiliaceae | 400 | 1400 | 400 | 1400 | ||
12 | Acer negundo L (AN) | Aceraceae | 400 | 1400 | 400 | 1400 | ||
13 | Acer mono Maxim (AM) | 400 | 1400 | 400 | 1400 | |||
14 | Larix gmelinii (Rupr.) Kuzen (LG) | Pinaceae | 400 | 1400 | 400 | 1400 | ||
15 | Picea koraiensis nakai (PK) | 400 | 1400 | 400 | 1400 | |||
16 | Abies balsamea (L.) Mill (AB) | 65 | 109 | |||||
17 | Abies nordmanniana Spach (AN) | 33 | 58 | |||||
Total | 4000 | 14000 | 838 | 4000 | 14000 | 1157 |
Based on the feature fusion of deep learning, ResNet50 [29] and DenseNet121 [30] are used as methods of tree species identification and cascade fusion, additive fusion and bilinear fusion are used as fusion algorithm. Before the fusion point, the same network model trains the bark and leaves separately, and fuse the feature of the leaves and bark at the fusion point. There are two fusion points, Fully connected layer and Convolution layer. The feature extraction network has ResNet50 and DenseNet121, and finally inputs them into softmax for recognition. In this fusion method, Fusion ResNet50 named FR50 and Fusion DenseNet121 named FD121, Fusion CNN named F-CNN.
Those models were trained on the Ubuntu 16.04 LTS 64 system on an GeForce GTX 1080Ti GPU hardware platform, equipped with Intel CoreTM i7-7800XCPU@3.50GHz×12 processors, Pytorch1.0 framework based on deep learning framework, using the Python and MATLAB language.
Features fusion belongs to the fusion of the intermediate level, and its basic principle is defined as the pattern space:
Ω={ω1,ω2,..,ωk} | (1) |
the sample set as X and Y, representing the bark and leaf. It is stated that Xj and Yj represent a feature set of X and Y, where j∈[1,p], p is the total number of samples for X and Y. The feature sets:
{Xj={X1j,X2j,…,XCj}Yj={Y1j,Y2j,…,YCj} | (2) |
Indicates that the sample has C features, and the sample feature vectors are:
{Xij={X1j1,X1j2,…,X1jni}∈XjYij={Y1j1,Y1j2,…,Y1jni}∈Yj | (3) |
X1jni x1jni and Y1jni represent one dimension of the i-th feature of the sample, and ni ni is the dimension of the feature. N=∑C1ni represents the total dimensions of the sample. Therefore, the correspondence between the sample feature sets Xj, Yj and ωk.
The two feature vectors output by the two convolution layers are fused to obtain the fused feature vectors, thus connecting the two CNN models together, the connection point is the fusion point. The fusion function is defined as:
f:x+y→L | (4) |
In Eq (4), x and y represent feature vectors of bark and leaf obtained by convolutional layer operations, L represents the fusion feature vectors of bark and leaf, and x, y, L∈RHWD, H、W、D respectively the length, width and number of channels [30].
Cascade fusion function is as follows:
Lcat=fcat(x,y) | (5) |
In Eq (5), the number of channels of the fused feature vectors is changed into two times of the original feature vector:
Lcati,j,2d=xi,j,d,Lcati,j,2d=yi,j,d | (6) |
In Eq (6), the L∈RH×W×2D.
Additive fusion is to add the values of the corresponding position elements of the two feature maps. The number of channels in the fused feature map is constant. Function is as follows:
Lsumi,j,d=xi,j,d+yi,j,d | (7) |
In Eq (7), the i∈[1,H],j∈[1,W],d∈[1,D],
Bilinear fusion is the summation of the position elements corresponding to two characteristic graphs after the outer product operation. The formula is:
Lbil=fbil(x,y) | (8) |
The number of channels in the fused feature map is the square of the number of channels in the original feature map, which is expressed as:
Lbil=∑Hi=1∑Wj=1xi,j⊗yi,j | (9) |
In Eq (9), the Lbil∈RD2. The fusion point is in the ReLu layer, and then the two feature images corresponding to the channel are fused.
Based on the fusion algorithm described above, we fuse leaves and bark on the ResNet50 and DenseNet121, the output of channel number of feature map is shown in the Table 2.
ReLu layer | Fully connected layer | ||
Bilinear fusion | Cascade fusion | Additive fusion | |
ResNet50 DenseNet121 |
1 × 1 × 512 → (1 × 1 × 512)2 1 × 1 × 1024 → (1 × 1 × 1024)2 |
1 × 1 × 512 → 2 × 1 × 1 × 512 1 × 1 × 1024 → 2 × 1 × 1 × 1024 |
1 × 1 × 512 → 1 × 1 × 512 1 × 1 × 1024 → 1 × 1 × 1024 |
In this paper, leaves and bark were combined as the input of F-CNN. As shown in Figure 3, there are two fusion points. One is bilinear fusion in the Convolution layer, which is input to the Fully connected layer, and then the classifier output. The second is cascade fusion and additive fusion in the Fully connected layer, then the classifier output.
The original dataset has 4000 images, divided into train set and test set, 3520 images and 880 images respectively. After flipping and translating, there are 28000 images, 22400 images as training set and 5600 images as test set. One is to use a single organ (bark or leaf) to classify and recognize in ResNet50 and DenseNet121 respectively. The other is to add the fusion algorithm to two network models. The third is to get the confusion matrix of single organ and two organ fusion respectively. The fourth is to compare the organ fusion method using CNN with single organ recognition, SVM and other fusion methods.
In order to make a fair comparison between the test results, the super parameters are standardized in the experiment, as Table 3.
Parameters | ResNet50 | DenseNet121 |
Batch size | 8 | 8 |
Epochs | 30 | 30 |
SGD momentum | 0.5 | 0.5 |
Weight decay | 0.005 | 0.005 |
Learning rate | 0.01 | 0.01 |
Cuda | Enable | Enable |
As shown in Table 4, it includes two network models, ResNet50 and DenseNet121, traditional classification algorithm SVM and other fusion algorithm. We give the recognition accuracy of each method. The first is single organ recognition, the recognition accuracy of bark is 75.75%, the recognition accuracy of leaves is 84.00%. The recognition accuracy of three fusion algorithms is 86.75%, 87.65%, 88.50%. For example, cascading fusion is 12.75% higher than bark recognition and 4.5% higher than leaf recognition by deep learning, 23.15% higher than bark recognition and 19.7% higher than leaf recognition by SVM, 12% higher than fusion recognition by SVM. Through the experimental verification, the fusion recognition of leaves and bark is better than single organ recognition and SVM recognition, which effectively solves the similarity and difference between species.
Original data | Fusion algorithms | ResNet50 | DenseNet121 | SVM |
Bark Leaf |
63.75 78.50 |
75.75 84.00 |
65.35 68.80 |
|
Bark + Leaf | Additive fusion | 83.50 | 86.75 | 74.30 |
Bark + Leaf | Bilinear fusion | 84.35 | 87.65 | 78.45 |
Bark + Leaf | Cascade fusion | 84.25 | 88.50 | 76.50 |
The following confusion matrix shows the effectiveness of the method used before the fusion, whether it has a good recognition effect on the image of leaves and bark, and what problems exist between trees, whether it will be recognized incorrectly and how many kinds of test sets are recognized correctly, and shows which classes are confused in the recognition process. In the matrix, the abscissa represents the prediction label and the ordinate represents the real label, diagonals represent the number of correctly predicted pictures. The darker the color is, the more correct pictures will be recognized. The following is the confusion matrix of ResNet50 and DenseNet121 of bark, leaf, the fusion of bark and leaf. As shown in Figure 4, the confusion matrix of bark, leaf, and fusion. Except for the first type of trees, the recognition accuracy of the remaining 9 kinds of trees has been improved. The confusion matrix shows the effectiveness of the proposed fusion algorithm for tree species identification.
Deep learning needs a large number of samples to form a suitable model. Due to the small amount of original data, the final recognition results are not very good, and it cannot objectively compare the fusion with a single organ, nor can it be shown that the fusion method is better than a single organ recognition, but also easy to be affected by over fitting. In order to solve the problem of sample number and over fitting, the data set is expanded by using the method of flipping and translation, in which the bark sample is expanded to 14000, and the leaf sample is expanded to 14000. Table 5 shows the recognition results after dataset expansion, the recognition accuracy of bark is 78.71%, the recognition accuracy of leaves is 86.75%. The recognition accuracy of three fusion algorithms is 93.02%, 90.43%, 93.17%.
Flipping and translation | Fusion algorithms | ResNet50 | DenseNet121 | SVM |
Bark Leaf |
71.64 86.75 |
78.71 86.50 |
70.34 68.96 |
|
Bark + Leaf | Additive fusion | 91.65 | 93.02 | 76.26 |
Bark + Leaf | Bilinear fusion | 90.43 | 89.48 | 78.85 |
Bark + Leaf | Cascade fusion | 92.42 | 93.17 | 77.91 |
Figure 5 is the confusion matrix of the extended dataset, which shows the single organ recognition and fusion recognition of each tree. Figure 5(a) is the confusion matrix under ResNet50 network, and Figure 5(b) is the confusion matrix under DenseNet121. Overall, more than half of the tree species recognition results improved.
The experiments of single organ recognition and fusion recognition are iterated 30 epochs, and the test results are shown in Figure 6. The red represents the recognition accuracy after the fusion of two organs (bark and leaf), the blue represents the recognition accuracy of leaf, and the green represents the recognition accuracy of bark. The red line is increasing and above the other two lines. The line graph also proves that the image fusion method based on bark and leaves has better effect on tree species recognition, and is obviously better than single organ recognition.
The images of leaf and bark of 7 kinds of trees were crawled through the network, as shown in Table 6. Through flipping and translation, the number of datasets is increased to obtain a more tree species recognition model to prevent overfitting due to data imbalance.
Species | Bark | Leaf | ||
Before expansion | After expansion | Before expansion | After expansion | |
Alnus glutinosa (L.) Gaertn | 221 | 440 | 234 | 440 |
Carya glabra (Mill.) Sweet | 35 | 210 | 76 | 210 |
Juglans regia L | 283 | 560 | 306 | 560 |
Quercus cerris L | 61 | 420 | 179 | 420 |
Fraxinus angustifolia Vahl | 140 | 540 | 195 | 540 |
Abies balsamea (L.) Mill | 65 | 455 | 109 | 455 |
Abies nordmanniana Spach | 33 | 210 | 58 | 210 |
We integrate the trees collected in the field with the trees crawled by the network, and train new CNN models. The classification results are shown in Table 7. The training form is that the training set is 80% of the total and the test set is 20% of the total. The training network is DenseNet121, the accuracy of bark identification is 61.91%, and the accuracy of leaf identification is 78.04%. The recognition accuracy after fusion is 87.86%. For example, [31] used comparable methods and achieved an accuracy of 74% for the combination of flower and leaf images using species from the PlantCLEF 2014 dataset. He et al.[32] extracted confidence scores for each single organ using a state-of-the-art DCNN, and deployed various schemes of the fusion approaches including not only conventional transformation-based approaches (sum rule, max rule, product rule), the accuracy of the proposed leaf and flower fusion technology reaches 82%. Therefore, it is proved that the recognition accuracy after fusion is better than single organ recognition.
Datasets | Fusion algorithms | ResNet50 | DenseNet121 | SVM |
Bark Leaf | 57.96 67.95 | 61.91 78.04 | 48.31 56.63 | |
Bark+Leaf | Additive fusion | 77.45 | 84.36 | 63.29 |
Bark+Leaf | Bilinear fusion | 78.32 | 85.67 | 65.46 |
Bark+Leaf | Cascade fusion | 78.77 | 87.86 | 66.71 |
Bark+Leaf | Other fusion algorithms [20] | 87 |
Figure 7 is the confusion matrix for all tree species. (a) and (b) are the confusion matrix of bark, leaf, and fusion respectively. In (a), the correct recognition numbers of the 1, 2, 3, 6, 7, 10, 13, 14 and 15 of bark are 22, 30, 32, 34, 51, 40, 37, 13, 13 and 25, the correct recognition numbers of the 1, 2, 3, 6, 7, 10, 13, 14, and 15 of leaf are 42, 28, 59, 32, 76, 34, 35, 26 and 39, the correct recognition numbers of the 1, 2, 3, 6, 7, 10, 13, 14, and 15 of fusion are and 73, 30, 88, 73, 98, 40, 40, 40 and 40. It also shows that the fusion method can solve the problem of intra class differences and inter class similarities better than single organ recognition.
Based on the confusion matrix and according to the classification of families, the recognition accuracy of each family is shown in Table 8. The fusion results of AG, JR, QC, FA, PA and TA are higher than that of single organ recognition. After fusion, the accuracy of Fagaceae, Oleaceae and Aceraceae was improved obviously. It can be further proved that the fusion can improve the accuracy of trees between intra class differences and inter class similarities.
Name | Tree species | Families | ResNet50 | DenseNet121 | ||||
Bark | Leaf | Fusion | Bark | Leaf | Fusion | |||
8 | BP | Betulaceae | 97.50 | 72.50 | 55.00 | 100.00 | 47.50 | 95.00 |
3 | AG | 36.36 | 67.05 | 100.00 | 25.00 | 80.68 | 92.05 | |
9 | JM | Juglandaceae | 100.00 | 7.41 | 43.75 | 97.50 | 19.05 | 68.75 |
4 | CG | 100.00 | 30.95 | 52.67 | 100.00 | 54.76 | 65.18 | |
5 | JR | 100.00 | 14.29 | 49.11 | 100.00 | 47.62 | 88.39 | |
6 | QC | Fagaceae | 40.47 | 38.09 | 86.91 | 72.50 | 66.67 | 92.85 |
10 | QM | 100.00 | 85.00 | 100.00 | 100.00 | 97.50 | 100.00 | |
11 | FM | Oleaceae | 90.00 | 87.50 | 75.00 | 95.00 | 97.50 | 90.00 |
7 | FA | 50.00 | 74.51 | 96.07 | 87.50 | 79.41 | 88.23 | |
12 | PA | Rutaceae | 75.00 | 77.50 | 85.00 | 92.50 | 82.50 | 100.00 |
13 | TA | Tiliaceae | 92.50 | 87.50 | 100.00 | 97.50 | 95.00 | 100.00 |
14 | AN | Aceraceae | 32.50 | 65.00 | 100.00 | 72.50 | 87.50 | 97.50 |
15 | AM | 62.50 | 97.50 | 100.00 | 55.00 | 100.00 | 100.00 | |
16 | LG | Pinaceae | 67.50 | 92.50 | 47.50 | 65.00 | 100.00 | 100.00 |
17 | PK | 86.50 | 95.00 | 72.50 | 90.00 | 90.00 | 100.00 | |
1 | AB | 24.17 | 46.15 | 80.22 | 39.56 | 68.13 | 37.36 | |
2 | AN | 71.43 | 66.67 | 71.43 | 100.00 | 59.52 | 97.62 |
For trees that are easily confused or difficult for people to differentiate, multiple organ recognition can be used to effectively improve. We used the fusion method of CNN and fusion algorithm to fuse leaves and bark. First, the networks of ResNet50 and DenseNet121 which are used to train the image of leaves and bark. Then the fusion algorithm is combined with CNN to realize the fusion of leaves and bark. The confusion matrix is used to illustrate and analyze the feasibility of the fusion method. Leaf and Bark provide quite different sources of information which, when used in combination, considerably improve the recognition result (Figures 4, 5 and 6). In addition, we also verified through three forms of data sets, one is the original dataset, which is 10 kinds of trees, the other is to expand on the basis of the original dataset, and the third is to increase the tree species, a total of 17 trees. The experimental results and confusion matrix show that the fusion method can improve the accuracy of tree species identification; it can effectively improve intra class differences and inter class similarities. This kind of fusion form also has reference significance for the recognition of other fields.
This work was supported by the Fundamental Research Funds for the Central Universities (Grants No. 2572017CB10), the Funding of Postdoctoral Research of Heilongjiang Province of China (Grants No. LBH-Z16006).
The author declares that he has no conflict of interest.
The image datasets used and analyzed during the current study are available from the corresponding author on reasonable request and under copyright restrictions.
[1] | P. Yuan, W. Li, S. Ren, H. Xu, Recognition for flower type and variety of chrysanthemum with convolutional neural network, Trans. Chin. Soc. Agric. Eng., 34 (2018), 152-158. |
[2] | G. Xuan, Y. Zhao, Q. Xiong, Z. Chen, Identification of Tree Species Based on Transfer Learning, For. Eng., 35 (2019), 68-75. |
[3] |
X. Zhu, M. Zhu, H. Ren, Method of plant leaf recognition based on improved deep convolutional neural network, Cognit. Syst. Res., 52 (2018), 223-233. doi: 10.1016/j.cogsys.2018.06.008
![]() |
[4] | H. Feng, M. Hu, Y. Yang, K. Xia, Tree Species Recognition Based on Overall Tree Image and Ensemble of Transfer Learning, Trans. Chin. Soc. Agric. Mach., 8 (2019), 235-279. |
[5] | L. Bi, Plant species recognition based on leaf image algorithm, Acta Agric. Zhejiangensis, 29 (2017), 2142-2148. |
[6] | P. Mittal, M. Kansal, H. K. Jhajj, Combined Classifier for Plant Classification and Identification from Leaf Image Based on Visual Attributes, 2018 International Conference on Intelligent Circuits and Systems, 2018. Available from: https://ieeexplore.ieee.org/abstract/document/8479567. |
[7] |
M. Kumar, S. Gupta, X. Gao, A. Singh, Plant Species Recognition Using Morphological Features and Adaptive Boosting Methodology, IEEE Access, 7 (2019), 163912-163918. doi: 10.1109/ACCESS.2019.2952176
![]() |
[8] |
N. Liu, J. Kan, Improved deep belief networks and multi-feature fusion for leaf identification, Neurocomputing, 216 (2016), 460-467. doi: 10.1016/j.neucom.2016.08.005
![]() |
[9] | X. Zhang, Y. Liu, H. Lin, Y. Liu, Research on SVM Plant Leaf Identification Method Based on CSA, International Conference of Pioneering Computer Scientists, Engineers and Educators, 2016. Available from: https://link.springer.com/chapter/10.1007/978-981-10-2098-8_20. |
[10] | T. P. Kumar, M. V. P. Reddy, P. K. Bora, Leaf Identification Using Shape and Texture Features, Proceedings of international conference on computer vision and image processing, 2017. Available from: https://link.springer.com/chapter/10.1007/978-981-10-2107-7_48. |
[11] | S. Boudra, I. Yahiaoui, A. Behloul, Statistical Radial Binary Patterns (SRBP) for Bark Texture Identification, International Conference on Advanced Concepts for Intelligent Vision Systems, 2017. Available from: https://link.springer.com/chapter/10.1007/978-3-319-70353-4_9. |
[12] | J. Chaki, R. Parekh, S. Bhattacharya, Plant leaf classification using multiple descriptors: A hierarchical approach, J. King Saud Univ., 2018 (2018). |
[13] | J. Chaki, N. Dey, L. Moraru, F. Shi, Fragmented plant leaf recognition: Bag-of-features, fuzzy-color and edge-texture histogram descriptors with multi-layer perceptron. Optik, 181 (2019), 639-650. |
[14] | S. L. Pimm, S. Alibhai, R. Berg, A. Dehgan, C. Giri, Z. Jewell, et al., Emerging technologies to conserve biodiversity. Trends. Ecol. Evol., 30 (2015), 685-696. |
[15] | M. Carpentier, P. Giguère, J. Gaudreault, Tree Species Identification from Bark Images Using Convolutional Neural Networks, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2018. Available from: https://ieeexplore.ieee.org/abstract/document/8593514/. |
[16] | J. Zheng, X. Wang, T. Wang, Research on image recognition of five bark texture images based on deep learning, J. Beijing For. Univ., 41 (2019), 150-158. |
[17] | S. Razavi, H. Yalcin, Using convolutional neural networks for plant classification, 2017 25th Signal Processing and Communications Applications Conference, 2017. Available from: https://ieeexplore.ieee.org/abstract/document/7960654. |
[18] | H. Zhou, C. Yan, H. Huang, Tree Species Identification Based on Convolutional Neural Network, 2016 8th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC), 2016. Available from: https://ieeexplore.ieee.org/abstract/document/7783797/. |
[19] | X. Zhang, J. Chen, J. ZhuGe, L. Yu, Deep Learning Based Fast Plant Image Recognition, J. East China Univ. Sci. Technol., 44 (2018), 105-113. |
[20] | H. Zhang, G. He, J. Peng, Z. Kuang, J. Fan, Deep Learning of Path-Based Tree Classifiers for Large-Scale Plant Species Identification, 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), 2018. Available from: https://ieeexplore.ieee.org/abstract/document/8396969. |
[21] | P. Guo, Q. Gao, A multi-organ plant identification method using convolutional neural networks, 2017 8th IEEE International Conference on Software Engineering and Service Science (ICSESS), 2017. Available from: https://ieeexplore.ieee.org/abstract/document/8342935. |
[22] |
S. Bertrand, R. B. Ameur, G. Cerutti, D. Coquin, L. Valet, L. Tougne, Bark and leaf fusion systems to improve automatic tree species recognition, Ecol. Inform., 46 (2018), 57-73. doi: 10.1016/j.ecoinf.2018.05.007
![]() |
[23] | M. Rzanny, P. Mä der, A. Deggelmann, M. Chen, J. Wä ldchen, Flowers, leaves or both? How to obtain suitable images for automated plant identification, Plant Methods, 15 (2019), 77. |
[24] | H. Goë au, P. Bonnet, A. Joly, V. Bakić, J. Barbe, I. Yahiaoui, et al., Pl@ntNet mobile app, Proceedings of the 21st ACM international conference on Multimedia, 2013. Available from: https://dl.acm.org/doi/abs/10.1145/2502081.2502251. |
[25] | iNaturalist, https://www.inaturalist.org/, Accessed 15 July 2019. |
[26] | J. Champ, T. Lorieul, M. Servajean, A. Joly, A comparative study of fine-grained classification methods in the context of the LifeCLEF plant identification challenge 2015, CLEF, 2015 (2015). |
[27] |
S. H. Lee, C. S. Chan, P. Remagnino, Multi-organ plant classification based on convolutional and recurrent neural networks, IEEE Trans. Image Process., 27 (2018), 4287-4301. doi: 10.1109/TIP.2018.2836321
![]() |
[28] | A. Joly, P. Bonnet, H. Goë au, J. Barbe, S. Selmi, J. Champ, et al., A look inside the pl@ntnet experience, Multimed Syst., 22 (2016), 751-766. |
[29] | K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, Proceedings of the IEEE conference on computer vision and pattern recognition, 2016. Available from: http://openaccess.thecvf.com/content_cvpr_2016/. |
[30] | G. Huang, Z. Liu, L. Maaten, K. Q. Weinberger, Densely connected convolutional networks, Proceedings of the IEEE conference on computer vision and pattern recognition, 2017. Available from: http://openaccess.thecvf.com/content_cvpr_2017/. |
[31] | W. B. Liu, Z. Y. Zou, W. W. Xing, Feature Fusion Methods in Pattern Classification, J. Beijing Univ. Posts Telecommun., 4 (2017), 5-12. |
[32] | A He, X Tian, Multi-organ plant identification with multi-column deep convolutional neural networks, 2016 IEEE international conference on systems, man, and cybernetics (SMC), 2016. Available from: https://ieeexplore.ieee.org/abstract/document/7844537. |
1. | Ryoungseob Kwon, Youngryel Ryu, Tackang Yang, Zilong Zhong, Jungho Im, Merging multiple sensing platforms and deep learning empowers individual tree mapping and species detection at the city scale, 2023, 206, 09242716, 201, 10.1016/j.isprsjprs.2023.11.011 | |
2. | Gokul Kottilapurath Surendran, Martin Lukac, Martin Lukac, Jozef Vybostok, Martin Mokros, A forestry investigation: Exploring factors behind improved tree species classification using bark images, 2024, 15749541, 102932, 10.1016/j.ecoinf.2024.102932 |
Name | Species | Families | Bark | Leaf | ||||
Before | After | Web crawling | Before | After | Web crawling | |||
1 | Betula platyphylla suk (BP) | Betulaceae | 400 | 1400 | 400 | 1400 | ||
2 | Alnus glutinosa (L.) Gaertn (AG) | 221 | 234 | |||||
3 | Juglans mandshurica maxim (JM) | Juglandaceae | 400 | 1400 | 400 | 1400 | ||
4 | Carya glabra (Mill.) Sweet (CG) | 35 | 76 | |||||
5 | Juglans regia L (JR) | 283 | 306 | |||||
6 | Quercus cerris L (QC) | Fagaceae | 61 | 179 | ||||
7 | Quercus mongolica fisch (QM) | 400 | 1400 | 400 | 1400 | |||
8 | Raxinus mandshurica rupr (FM) | Oleaceae | 400 | 1400 | 400 | 1400 | ||
9 | Fraxinus angustifolia Vahl (FA) | 140 | 195 | |||||
10 | Phellodendron amurense Rupr (PA) | Rutaceae | 400 | 1400 | 400 | 1400 | ||
11 | Tilia amurensis Rupr (TA) | Tiliaceae | 400 | 1400 | 400 | 1400 | ||
12 | Acer negundo L (AN) | Aceraceae | 400 | 1400 | 400 | 1400 | ||
13 | Acer mono Maxim (AM) | 400 | 1400 | 400 | 1400 | |||
14 | Larix gmelinii (Rupr.) Kuzen (LG) | Pinaceae | 400 | 1400 | 400 | 1400 | ||
15 | Picea koraiensis nakai (PK) | 400 | 1400 | 400 | 1400 | |||
16 | Abies balsamea (L.) Mill (AB) | 65 | 109 | |||||
17 | Abies nordmanniana Spach (AN) | 33 | 58 | |||||
Total | 4000 | 14000 | 838 | 4000 | 14000 | 1157 |
ReLu layer | Fully connected layer | ||
Bilinear fusion | Cascade fusion | Additive fusion | |
ResNet50 DenseNet121 |
1 × 1 × 512 → (1 × 1 × 512)2 1 × 1 × 1024 → (1 × 1 × 1024)2 |
1 × 1 × 512 → 2 × 1 × 1 × 512 1 × 1 × 1024 → 2 × 1 × 1 × 1024 |
1 × 1 × 512 → 1 × 1 × 512 1 × 1 × 1024 → 1 × 1 × 1024 |
Parameters | ResNet50 | DenseNet121 |
Batch size | 8 | 8 |
Epochs | 30 | 30 |
SGD momentum | 0.5 | 0.5 |
Weight decay | 0.005 | 0.005 |
Learning rate | 0.01 | 0.01 |
Cuda | Enable | Enable |
Original data | Fusion algorithms | ResNet50 | DenseNet121 | SVM |
Bark Leaf |
63.75 78.50 |
75.75 84.00 |
65.35 68.80 |
|
Bark + Leaf | Additive fusion | 83.50 | 86.75 | 74.30 |
Bark + Leaf | Bilinear fusion | 84.35 | 87.65 | 78.45 |
Bark + Leaf | Cascade fusion | 84.25 | 88.50 | 76.50 |
Flipping and translation | Fusion algorithms | ResNet50 | DenseNet121 | SVM |
Bark Leaf |
71.64 86.75 |
78.71 86.50 |
70.34 68.96 |
|
Bark + Leaf | Additive fusion | 91.65 | 93.02 | 76.26 |
Bark + Leaf | Bilinear fusion | 90.43 | 89.48 | 78.85 |
Bark + Leaf | Cascade fusion | 92.42 | 93.17 | 77.91 |
Species | Bark | Leaf | ||
Before expansion | After expansion | Before expansion | After expansion | |
Alnus glutinosa (L.) Gaertn | 221 | 440 | 234 | 440 |
Carya glabra (Mill.) Sweet | 35 | 210 | 76 | 210 |
Juglans regia L | 283 | 560 | 306 | 560 |
Quercus cerris L | 61 | 420 | 179 | 420 |
Fraxinus angustifolia Vahl | 140 | 540 | 195 | 540 |
Abies balsamea (L.) Mill | 65 | 455 | 109 | 455 |
Abies nordmanniana Spach | 33 | 210 | 58 | 210 |
Datasets | Fusion algorithms | ResNet50 | DenseNet121 | SVM |
Bark Leaf | 57.96 67.95 | 61.91 78.04 | 48.31 56.63 | |
Bark+Leaf | Additive fusion | 77.45 | 84.36 | 63.29 |
Bark+Leaf | Bilinear fusion | 78.32 | 85.67 | 65.46 |
Bark+Leaf | Cascade fusion | 78.77 | 87.86 | 66.71 |
Bark+Leaf | Other fusion algorithms [20] | 87 |
Name | Tree species | Families | ResNet50 | DenseNet121 | ||||
Bark | Leaf | Fusion | Bark | Leaf | Fusion | |||
8 | BP | Betulaceae | 97.50 | 72.50 | 55.00 | 100.00 | 47.50 | 95.00 |
3 | AG | 36.36 | 67.05 | 100.00 | 25.00 | 80.68 | 92.05 | |
9 | JM | Juglandaceae | 100.00 | 7.41 | 43.75 | 97.50 | 19.05 | 68.75 |
4 | CG | 100.00 | 30.95 | 52.67 | 100.00 | 54.76 | 65.18 | |
5 | JR | 100.00 | 14.29 | 49.11 | 100.00 | 47.62 | 88.39 | |
6 | QC | Fagaceae | 40.47 | 38.09 | 86.91 | 72.50 | 66.67 | 92.85 |
10 | QM | 100.00 | 85.00 | 100.00 | 100.00 | 97.50 | 100.00 | |
11 | FM | Oleaceae | 90.00 | 87.50 | 75.00 | 95.00 | 97.50 | 90.00 |
7 | FA | 50.00 | 74.51 | 96.07 | 87.50 | 79.41 | 88.23 | |
12 | PA | Rutaceae | 75.00 | 77.50 | 85.00 | 92.50 | 82.50 | 100.00 |
13 | TA | Tiliaceae | 92.50 | 87.50 | 100.00 | 97.50 | 95.00 | 100.00 |
14 | AN | Aceraceae | 32.50 | 65.00 | 100.00 | 72.50 | 87.50 | 97.50 |
15 | AM | 62.50 | 97.50 | 100.00 | 55.00 | 100.00 | 100.00 | |
16 | LG | Pinaceae | 67.50 | 92.50 | 47.50 | 65.00 | 100.00 | 100.00 |
17 | PK | 86.50 | 95.00 | 72.50 | 90.00 | 90.00 | 100.00 | |
1 | AB | 24.17 | 46.15 | 80.22 | 39.56 | 68.13 | 37.36 | |
2 | AN | 71.43 | 66.67 | 71.43 | 100.00 | 59.52 | 97.62 |
Name | Species | Families | Bark | Leaf | ||||
Before | After | Web crawling | Before | After | Web crawling | |||
1 | Betula platyphylla suk (BP) | Betulaceae | 400 | 1400 | 400 | 1400 | ||
2 | Alnus glutinosa (L.) Gaertn (AG) | 221 | 234 | |||||
3 | Juglans mandshurica maxim (JM) | Juglandaceae | 400 | 1400 | 400 | 1400 | ||
4 | Carya glabra (Mill.) Sweet (CG) | 35 | 76 | |||||
5 | Juglans regia L (JR) | 283 | 306 | |||||
6 | Quercus cerris L (QC) | Fagaceae | 61 | 179 | ||||
7 | Quercus mongolica fisch (QM) | 400 | 1400 | 400 | 1400 | |||
8 | Raxinus mandshurica rupr (FM) | Oleaceae | 400 | 1400 | 400 | 1400 | ||
9 | Fraxinus angustifolia Vahl (FA) | 140 | 195 | |||||
10 | Phellodendron amurense Rupr (PA) | Rutaceae | 400 | 1400 | 400 | 1400 | ||
11 | Tilia amurensis Rupr (TA) | Tiliaceae | 400 | 1400 | 400 | 1400 | ||
12 | Acer negundo L (AN) | Aceraceae | 400 | 1400 | 400 | 1400 | ||
13 | Acer mono Maxim (AM) | 400 | 1400 | 400 | 1400 | |||
14 | Larix gmelinii (Rupr.) Kuzen (LG) | Pinaceae | 400 | 1400 | 400 | 1400 | ||
15 | Picea koraiensis nakai (PK) | 400 | 1400 | 400 | 1400 | |||
16 | Abies balsamea (L.) Mill (AB) | 65 | 109 | |||||
17 | Abies nordmanniana Spach (AN) | 33 | 58 | |||||
Total | 4000 | 14000 | 838 | 4000 | 14000 | 1157 |
ReLu layer | Fully connected layer | ||
Bilinear fusion | Cascade fusion | Additive fusion | |
ResNet50 DenseNet121 |
1 × 1 × 512 → (1 × 1 × 512)2 1 × 1 × 1024 → (1 × 1 × 1024)2 |
1 × 1 × 512 → 2 × 1 × 1 × 512 1 × 1 × 1024 → 2 × 1 × 1 × 1024 |
1 × 1 × 512 → 1 × 1 × 512 1 × 1 × 1024 → 1 × 1 × 1024 |
Parameters | ResNet50 | DenseNet121 |
Batch size | 8 | 8 |
Epochs | 30 | 30 |
SGD momentum | 0.5 | 0.5 |
Weight decay | 0.005 | 0.005 |
Learning rate | 0.01 | 0.01 |
Cuda | Enable | Enable |
Original data | Fusion algorithms | ResNet50 | DenseNet121 | SVM |
Bark Leaf |
63.75 78.50 |
75.75 84.00 |
65.35 68.80 |
|
Bark + Leaf | Additive fusion | 83.50 | 86.75 | 74.30 |
Bark + Leaf | Bilinear fusion | 84.35 | 87.65 | 78.45 |
Bark + Leaf | Cascade fusion | 84.25 | 88.50 | 76.50 |
Flipping and translation | Fusion algorithms | ResNet50 | DenseNet121 | SVM |
Bark Leaf |
71.64 86.75 |
78.71 86.50 |
70.34 68.96 |
|
Bark + Leaf | Additive fusion | 91.65 | 93.02 | 76.26 |
Bark + Leaf | Bilinear fusion | 90.43 | 89.48 | 78.85 |
Bark + Leaf | Cascade fusion | 92.42 | 93.17 | 77.91 |
Species | Bark | Leaf | ||
Before expansion | After expansion | Before expansion | After expansion | |
Alnus glutinosa (L.) Gaertn | 221 | 440 | 234 | 440 |
Carya glabra (Mill.) Sweet | 35 | 210 | 76 | 210 |
Juglans regia L | 283 | 560 | 306 | 560 |
Quercus cerris L | 61 | 420 | 179 | 420 |
Fraxinus angustifolia Vahl | 140 | 540 | 195 | 540 |
Abies balsamea (L.) Mill | 65 | 455 | 109 | 455 |
Abies nordmanniana Spach | 33 | 210 | 58 | 210 |
Datasets | Fusion algorithms | ResNet50 | DenseNet121 | SVM |
Bark Leaf | 57.96 67.95 | 61.91 78.04 | 48.31 56.63 | |
Bark+Leaf | Additive fusion | 77.45 | 84.36 | 63.29 |
Bark+Leaf | Bilinear fusion | 78.32 | 85.67 | 65.46 |
Bark+Leaf | Cascade fusion | 78.77 | 87.86 | 66.71 |
Bark+Leaf | Other fusion algorithms [20] | 87 |
Name | Tree species | Families | ResNet50 | DenseNet121 | ||||
Bark | Leaf | Fusion | Bark | Leaf | Fusion | |||
8 | BP | Betulaceae | 97.50 | 72.50 | 55.00 | 100.00 | 47.50 | 95.00 |
3 | AG | 36.36 | 67.05 | 100.00 | 25.00 | 80.68 | 92.05 | |
9 | JM | Juglandaceae | 100.00 | 7.41 | 43.75 | 97.50 | 19.05 | 68.75 |
4 | CG | 100.00 | 30.95 | 52.67 | 100.00 | 54.76 | 65.18 | |
5 | JR | 100.00 | 14.29 | 49.11 | 100.00 | 47.62 | 88.39 | |
6 | QC | Fagaceae | 40.47 | 38.09 | 86.91 | 72.50 | 66.67 | 92.85 |
10 | QM | 100.00 | 85.00 | 100.00 | 100.00 | 97.50 | 100.00 | |
11 | FM | Oleaceae | 90.00 | 87.50 | 75.00 | 95.00 | 97.50 | 90.00 |
7 | FA | 50.00 | 74.51 | 96.07 | 87.50 | 79.41 | 88.23 | |
12 | PA | Rutaceae | 75.00 | 77.50 | 85.00 | 92.50 | 82.50 | 100.00 |
13 | TA | Tiliaceae | 92.50 | 87.50 | 100.00 | 97.50 | 95.00 | 100.00 |
14 | AN | Aceraceae | 32.50 | 65.00 | 100.00 | 72.50 | 87.50 | 97.50 |
15 | AM | 62.50 | 97.50 | 100.00 | 55.00 | 100.00 | 100.00 | |
16 | LG | Pinaceae | 67.50 | 92.50 | 47.50 | 65.00 | 100.00 | 100.00 |
17 | PK | 86.50 | 95.00 | 72.50 | 90.00 | 90.00 | 100.00 | |
1 | AB | 24.17 | 46.15 | 80.22 | 39.56 | 68.13 | 37.36 | |
2 | AN | 71.43 | 66.67 | 71.43 | 100.00 | 59.52 | 97.62 |