Abstract
The significance of automatic plant identification has already been recognized by academia and industry. There were several attempts to utilize leaves and flowers for identification; however, bark also could be beneficial, especially for trees, due to its consistency throughout the seasons and its easy accessibility, even in high crown conditions. Previous studies regarding bark identification have mostly contributed quantitatively to increasing classification accuracy. However, ever since computer vision algorithms surpassed the identification ability of humans, an open question arises as to how machines successfully interpret and unravel the complicated patterns of barks. Here, we trained two convolutional neural networks (CNNs) with distinct architectures using a large-scale bark image dataset and applied class activation mapping (CAM) aggregation to investigate diagnostic keys for identifying each species. CNNs could identify the barks of 42 species with > 90% accuracy, and the overall accuracies showed a small difference between the two models. Diagnostic keys matched with salient shapes, which were also easily recognized by human eyes, and were typified as blisters, horizontal and vertical stripes, lenticels of various shapes, and vertical crevices and clefts. The two models exhibited disparate quality in the diagnostic features: the old and less complex model showed more general and well-matching patterns, while the better-performing model with much deeper layers indicated local patterns less relevant to barks. CNNs were also capable of predicting untrained species by 41.98% and 48.67% within the correct genus and family, respectively. Our methodologies and findings are potentially applicable to identify and visualize crucial traits of other plant organs.
Similar content being viewed by others
Introduction
Species identification is a fundamental component in every discipline of biology to properly utilize, monitor, and protect highly diverse living organisms on Earth. However, conventional identification workflows that manually distinguish key visual features are slow and error-prone because of the complexity and intraspecific variation in morphological traits1. There is also increasing concern that the number of professional and amateur taxonomists is persistently declining, and that the gap in taxonomic knowledge between professionals and the public is increasing1,2,3. Thus, an increasing number of studies propose automated identification systems based on rapidly advancing machine learning methods, to support both professionals and the public4,5,6,7,8. Among the various fields where automation is being introduced, plant identification is one of the most vigorously studied areas with high demand owing to its easy accessibility, rich diversity, and increasing curiosity about natural creatures in urban life9.
In addition to conventional taxonomic approaches to identify plants, previous studies regarding automated plant identification have mainly focused on extracting visual features from either or both reproductive and vegetative organs10,11,12,13,14,15. However, the availability of these organs highly depends on the season, phenological changes, and height of the crown base, especially in tree species. Hence, owing to the non-rigid and three-dimensional nature of leaves and flowers, it is challenging to produce high-quality image data to train and evaluate machine learning models16. The challenges encountered in identification by leaves and flowers, have led several studies to exploit the benefit of using barks that persist throughout the season and are easy to access and process with their simple cylindrical shapes17.
Limited studies on bark identification have focused on the terminology and description of visual patterns in the inner and outer bark18. Whitmore19 and Yunus, et al.20 visually inspected tropical trees and broadly categorized the patterns (e.g., deep/shallow fissured, scaly, and laminate). Prior to technological breakthroughs, taxonomists utilized these little-known bark traits as an auxiliary means of species identification. Recently, Hadlich, et al.21 used a portable spectrometer to identify bark from tropical species. Other methodologies, utilizing Red, Green, Blue (RGB) channel images, approached the problem as multi-class texture classification task consisted of two consecutive phases: representation and classification22. Representation (feature extraction) is an abstraction process that extracts useful features from the labyrinthine texture of barks, while classification assigns the features into discrete class labels by minimizing the error function of a specific machine learning algorithm. Early attempts mainly took advantage of hand-crafted features, including Gabor filters23, SIFT24, local binary patterns25, and local ternary patterns22, which were followed by classification using conventional machine learning algorithms (e.g., k-nearest neighbors, support vector machines, and multi-layer perceptron). Recently, these human-designed feature representation workflows have been rapidly substituted by data-driven approaches in which machines find the most appropriate features. Regarding data-driven approaches, especially for large-scale image data, convolutional neural networks (CNNs) are the most commonly adopted algorithms that are superior to conventional ones26. Carpentier, et al.27 also showed that CNNs could identify 23 different tree species with > 90% accuracy using only small-patched images of barks, which is a highly challenging task even for professionals24.
Despite the powerful performance of CNNs, there is an inevitable trade-off between the accuracy and interpretability of the model28,29. Due to the complexity and the large number of parameters of CNNs that reach over 480 million in current state-of-the-art approaches30,31, it is almost impossible to review the individual parameters by tracing back the model’s output. Lack of interpretability and the ability to explain the model’s prediction and accuracy could lead to concerns regarding the legitimacy and mystery of the performance of CNNs. These concerns are intensified by its weak taxonomic foundation and fewer morphological characteristics32. Previous studies also mainly focused on the contribution to the quantitative aspects by raising accuracy in specific datasets rather than expanding qualitative knowledge, which can be useful for professionals in other disciplines.
In this aspect, we focus on discovering the qualitative features of how state-of-the-art computer vision machines successfully unravel highly sophisticated patterns of nature. We used the latest advances in computer vision, class activation mapping (CAM), which enabled us to access one of the nested groups of parameters (layers) in CNNs and interpret the output by highlighting the most influential (activated) regions in making the prediction29,33,34. We prepared a bark image dataset of 42 tree species and trained two well-characterized CNN architectures: the old, simple, and effective VGG-16 and the state-of-the-art, but with high architectural complexity, EfficientNet. After the models were successfully trained and validated, we extracted and generated saliency maps for the entire sample with respect to each model. We proposed a novel approach called CAM aggregation (“Materials and methods”; Fig. 7), that enables the unified representation of diagnostic features over the entire bark. The heat maps were manually reviewed and documented as diagnostic keys for each species, and the reliability and limitations of our proposed method were investigated.
More specifically, we aimed to answer the following research questions:
(1) Can machine-learning algorithms be applied to identify woody species using bark characteristics? What accuracy can be expected from different samples?
(2) How do two machine-learning CNN algorithms (VGG-16 and EfficientNet) compare?
(3) Does the interpretation methodology (i.e., CAM) reveal specific diagnostic features that are important for their successful application?
Results
The training of the two CNN models was successful, and both CNN models reached over 90% overall accuracy, as shown in Table 1. The performance metrics (“Materials and methods”) did not show substantial differences among the four measures, even between the models (ranged within 0.5%). A more complex model, namely EfficientNet, required a much larger amount of training to reach its maximum accuracy; it took 65 epochs in VGG-16 (90.7%) and 161 epochs in EfficientNet (91.0%). A more detailed comparison of classification performance with respect to each species is given as confusion matrices in Figs. 1 and 2, produced by VGG-16 and EfficientNet, respectively. Most species showed high accuracies of > 90%, but two models tended to show lower accuracies in some species that had multiple species in each genus or family, and they were mostly misclassified as other species in the same genus or family.
In VGG-16, for example, large misclassifications were made in Abies balsamea (by 6.39% to Picea mariana), Castanea crenata (by 7.0% and 6.6% to Quercus aliena and Quercus serrata, respectively), Taxodium distichum (by 12.7% to Metasequoia glyptostroboides), Acer rubrum (by 11.5% to Acer saccharum), Picea glauca (by 18.7% to Picea abies), and Quercus serrata (by 33.9% to Quercus aliena). EfficientNet followed the same trend but showed weaknesses in slightly different species: Abies balsamea (by 8.0% to Picea mariana), Acer rubrum (by 14.5% to Acer saccharum), Betula alleghaniensis (by 14.47% to Betula papyrifera), and Quercus aliena (by 12.6% to Quercus serrata). Ulmus americana was misclassified to species in other taxon groups, including Acer saccharum (9.5%) and Ginkgo biloba (5.1%).
Meanwhile, the accuracy of the CNN models for each species did not show a clear trend with the number of training samples (Figs. 1 and 2, and Table 2). Species with high error rates (e.g., Acer rubrum, Castanea crenata, Picea glauca, and Quercus serrata) were trained with an intermediate number of unique bark images, ranging from 161 to 226, whereas species trained with a relatively small number of images still showed high accuracies (e.g., Acer palmatum var. amoenum, Magnolia obovata, and Picea rubens). However, considering that the ImageNet dataset consisted of approximately 1000 images per class35, we suggest that more than 100 images per species should be prepared, assuming that 10 cropped pieces can be sampled for each bark image (Fig. 6, “Data pre-processing and augmentation”).
These results suggest that CNNs can successfully identify most species; however, varying the extent of intra-genus and intra-family similarities could lead to low accuracy, and two CNNs might learn and utilize classification keys in different ways.
The diagnostic features of each species were identified as heat maps displaying the relative importance of classification (Fig. 3 and extended figures at https://doi.org/10.6084/m9.figshare.14727834). In the heat maps extracted from VGG-16, the saliency responses in most species were typically matched with localized shape patterns, which were also readily noticeable by human eyes. For example, the patterns were shown as blisters in Abies, vertical stripes and crevices in Acer, large horizontal lenticels in Betula, shaded parts of scaly bark in Picea, scaly and flaking patterns in Platanus, small striped and dotted lenticels in Prunus, and vertical crevices and clefts in Quercus (detailed information by species listed in SI Appendix, Table S1). A large number of images was activated by higher density in several species that did not retain salient features but exhibited repeated textures along the entire bark. It was mostly found in four Cupressaceae species (Chamaecyparis pisifera, Metasequoia glyptostroboides, Taxodium distichum, and Thuja occidentalis) and in some samples of Fagus grandifolia, and Magnolia obovata, as large numbers of elongated narrow cracks or stripes. These results indicate that VGG-16 mostly catches locally prominent features if available, and if not, they utilize the entire texture patterns that are extensively shown and repeated.
In EfficientNet, the disparity of heat maps in terms of both quantity and quality was observed. The model utilized much smaller and weakly activated regions that did not match the proposed patterns from VGG-16 in many species. Along with smaller sizes of each activation region, they were less matched with the exact location of salient features including stripes, lenticels, and crevices, and also showed activation for some digressive and irrelevant features of barks (e.g., backgrounds, mosses, knots, leaves, and catkins) (Fig. 3 and extended figures at https://doi.org/10.6084/m9.figshare.14727834).
Discussion
Our results suggest that CNNs have sufficient capacity to identify tree bark images with high accuracy in most species. However, the proposed methods also showed weaknesses in distinguishing the inter-species similarities between some genera and families, which might be attributed to disparity in diagnostic features, whereas the overall accuracy did not show substantial differences between the two models.
The contrasting diagnostic features from EfficientNet could be explained by its complicated architecture consisting of 54 convolution layers, 17 pooling layers, and other dropout and reshape layers (240 layers in total). In contrast, VGG-16 consists of 13 convolution layers and five pooling layers (23 layers in total). Convolution layers at different depths learn distinct levels of representation, and the deepest layers are trained to capture the features with the highest abstraction36. It could be inferred that increasing the number of convolution layers and architectural complexity led the CNN model to utilize more condensed information in the data.
In this aspect, augmenting the complexity of the model increased performances of large-scale image classification datasets (e.g., ImageNet) of objects with rigid boundaries and intermediate intra-class variances. However, natural images show large variation in the same species (age and environmental variation) and frequently show high inter-species similarity. If the model excessively focuses on the local features shown in a specific entity rather than grasping general patterns shared at the species level, it might overfit and decrease overall performance.
Previous studies reported that ImageNet pre-trained models showed high domain adaptation ability. That is, high accuracy models on ImageNet showed higher accuracies on other domains (R2 > 0.95), tested with a benchmark dataset consisted of 65 objects (Office-Home Dataset)37. In this study, EfficientNet did not show substantially better accuracy than VGG-16, regardless of the superior performance on ImageNet.
The varying accuracies and CAM representation capacities of CNNs with natural images have been studied less, but few studies are in accordance with the findings of this study. When tested for detecting flowering events using five distinctive CNN architectures38, the accuracies ranged only within 1.3%, but CAM visualization yielded largely contrasting results. In the study, simpler models showed activation to the large and relevant region to flowers, whereas the complex and better performing model showed much smaller activated region. Therefore, the complex model predicted flowering by only a small number of flowers with higher sensitivity.
In the case of barks, augmented complexity and sensitivity could have mixed effects on both positive and negative sides. Complex models might identify species with more localized and sophisticated features that simple models cannot find, but along with the risk of overfitting. Consequently, ensemble techniques that combine and predict different types of models or architectural modifications that enable multi-receptive levels (e.g., Inception modules in GoogLeNet39) would be suitable for real-world applications.
A qualitative inspection of misclassified samples in Fig. 1 and 2 revealed that CNNs tended to show errors in some typical conditions which should be avoided in data preparation. They could be typified as excessive background masking in Fig. 4a, foreign object coverage in Fig. 4b, and improper light conditions in Fig. 4c. The other misclassifications that could not be typified, may originated from high inter-species similarity along with the black-box characteristics of deep learning models. As shown by Rzanny, et al.14, where fusing multiple organs increased accuracy up to 20%, combining the bark with leaf and flower characteristics would lead to improved accuracy and robustness.
Meanwhile, CNN models tend to misclassify other species in the same genus or family (Figs. 1 and 2). These results imply that diagnostic features are shared within higher taxa, and CNNs can recognize and utilize these general features for identification. We further investigated this generalization capacity with images of new species within the same genus or family27, consisting of 20 species in Abies, Acer, Alnus, Betula, Carpinus, Castanea, Chionanthus, Fagus, Fraxinus, Platanus, Populus, Quercus, and Ulmus (Fig. 5). When tested with two models, VGG-16 predicted 42.0% of the total samples to the species in the same genus (among 42 original species), and 48.7% to the species in the same family; and EfficientNet predicted 37.8% and 44.5% of the species in the same genus and family, respectively. In detail, VGG-16 showed high generalization ability in Acer pictum subsp. mono (93.1%) and Acer psuedosieboldianum (68.6%) into Acer palmatum var. amoenum, Platanus orientalis (49.2%) into Platanus occidentalis, Quercus dentata (98.1%) and Quercus × urticaefolia (100%) into Quercus aliena. EfficientNet particularly showed better results in Betula platyphylla (45.3%), Fagus sylvatica (53.3%), Platanus orientalis (82.0%) and Populus grandidentata (66.4%). These results suggest that CNNs can catch the shared features within a higher taxon and categorize new species into their genera or families if trained before. Furthermore, it seems that VGG-16 learned and utilized more generalized features from the original dataset than EfficientNet, which showed more localized and digressive diagnostic feature representations (Fig. 3 and extended figures at https://doi.org/10.6084/m9.figshare.14727834).
In summary, we demonstrated the capacity of CNNs to distinguish sophisticated patterns in barks and proposed a novel approach to extract and narrate diagnostic keys found by computer vision models. We insist that architectural differences in CNNs largely affect the quality of diagnostic features, which is not shown by quantitative accuracy metrics. Our proposed methods are easily applicable to other plant organs (e.g., leaves, flowers, and wood anatomical features) that also exhibit complex patterns on their surfaces. Furthermore, we suggest that using the bark as a baseline prediction factor and combining the information from other organs which will enable highly precise tree species identification.
Materials and methods
Bark image datasets
The Bark dataset for training and evaluation of CNNs was prepared by combining two different datasets: BarkNet 1.027 and the additional dataset collected for this study. In terms of size and number of species, BarkNet is currently the only publicly available dataset of its kind and contains more than 23,000 bark images captured from 23 tree species found in Quebec City, Canada, with 10–40 repeated captures from different angles and heights of individual trees. The captured raw images were pre-processed in the dataset by two discrete crop processes: trimming off two vertical border lines between the trunk and the background (side-crop) and then cropping into square-shaped patches to match the input shape for CNNs (square-crop). In this study, side-cropped images of the BarkNet dataset were utilized. As the images in BarkNet were captured with a large number of repetitions, this might lead to high train–test dataset similarity, which impedes efficient training of CNNs. Consequently, four images were randomly sampled from each of 10 to 40 repetitions and then manually filtered out of images that share a large portion of each other.
For the additional dataset, the images were captured according to the protocols of BarkNet with minor modifications. The protocols were as follows: captured from September to November 2020; in parks and forests near Seoul City, Republic of Korea; with a wide range of crown density; under various weather conditions from sunny to cloudy and rainy; using an iPhone SE2 camera; at a distance between 20 and 60 cm away from the trunk; with three to four repetitions from different angles per individual tree. The species selected were common species that can be easily found in most parks and forests in the Republic of Korea. We collected samples from multiple species in each genus to investigate intra-genus similarities.
Consequently, two datasets consisting of 3672 images (20 species; BarkNet 1.0) and 5703 images (23 species; for the present study) were prepared, having one species Pinus strobus in common. A more detailed list of species and the corresponding number of captured images in the two datasets are given in Table 2.
Data pre-processing and augmentation
The overall pre-processing and augmentation are illustrated in Fig. 6. Before aggregating the two datasets, side-crop and masking were applied to our dataset. Because our dataset contained images that had non-vertical boundary lines, all images were side-cropped to a length that did not cut out pixels of barks, and the remaining background was masked by black pixels. After the images were cropped and masked, BarkNet and our dataset were aggregated and split into two subsets along with each class: 80% and 20% for training and testing, respectively.
For the training dataset, a random square crop with thresholding was applied to the images before putting them into the CNNs. Unlike previous studies, the images were randomly square-cropped by proportional length instead of fixed pixel lengths, such as 224 × 22427 and 256 × 25640. For each crop, the pixel length of the crop was first randomly selected in the range of 40–60% of the total width, and then the position of the cropping square was randomly sampled. To exclude excessive background, the ratio of masked pixels to total pixels was calculated, and if the ratio was higher than the threshold (0.05, in this study), the crop region was re-sampled. This approach has several advantages: minimizing the loss of bark features that overlap with crop region, flexible cropping that copes with varying resolutions, and cropped images that require no manual filtering.
For data augmentation, a process that increases the training data by giving slight modifications, ‘RandAugment’ was utilized41. RandAugment is an automated data augmentation process that randomly selects a predefined number of augmentation functions and then applies them with a randomly sampled magnitude from a predefined range. The ‘candidate’ functions include equalization, rotation, solarization, adjustment of contrast, brightness, and color. The entire list of augmentation functions and hyperparameters regarding RandAugment could be found in the provided Python scripts (“Materials and methods”: Data and code availability).
These random square-crop and augmentation processes were applied concurrently with the training. More specifically, side-cropped training samples remained unmodified, the random square-crop followed by augmentation was performed, and the output images with the number of predefined batch sizes were fed into the CNNs. For the testing dataset, each side-cropped sample was square-cropped by 50% of the total width, without randomized sampling, and no augmentation was applied. All training and testing samples were resized to 331 × 331 before data augmentation.
Convolution neural networks
In this study, among numerous CNN architectures that have been extensively studied and released, the following architectures were adopted and compared: VGG-16 and EfficientNet42,43. As one of the early CNN architectures released in 2014, VGG-16 remains a widely used CNN architecture due to its simple structure and powerful performance in various computer vision tasks, including classification44, detection45, and segmentation46. In contrast, EfficientNet is currently the most powerful CNN architecture, showing the highest benchmark accuracy on ImageNet (26) while reducing the number of parameters compared to other recent models. EfficientNet is an approach that searches for the best scaling coefficients in terms of the width, depth, and resolution of CNNs. In this study, EfficientNet with the smallest scale size (B0) was used to match the input size with VGG-16.
Using two CNN architectures, each model was modified for effective transfer learning by following sequences: (1) truncate the final fully-connected (FC) layer which contained neurons for predicting 1000 classes (in the original ImageNet dataset), (2) reshape (flatten) the output of the final convolution layer and apply dropout, (3) insert a new FC layer with the same number of neurons as the prediction classes (42 species), (4) apply softmax activation with cross-entropy loss functions.
Here, a convolution layer performs element-wise multiplication between the input data and the weights; a pooling layer simply reduces the dimensions of the input data; a dropout layer randomly sets input node values to zero and prevents overfitting; and a FC layer connects all nodes in one layer to every node in itself.
Extracting diagnostic features with class activation mapping aggregation
Since CAM was first proposed by Zhou, et al.33, there have been several advances in CAM techniques. The original CAM had certain limitations in that it required a global average pooling layer in the architecture, along with laborious re-training. These limitations were overcome by Selvaraju, et al.29 in gradient-weighted class activation mapping (Grad-CAM), using a back-propagation gradient in any target convolution layer. An advanced method called Grad-CAM++ has recently been proposed, which provides better localization capability than Grad-CAM, especially with multiple object instances occurring in a single image. Considering that the diagnostic features would show multiple occurrences in a single bark image, Grad-CAM++ was utilized in this study as the CAM extraction method.
To build a general representation of diagnostic features over the entire bark rather than individual square-cropped patches, aggregation and averaging methods were applied, referred to as CAM aggregation (Fig. 7). The CAM of each tree image was calculated using the following steps: zero-padding, CAM calculation from sliding windows, and taking the mean of CAM outputs. More specifically, bark images of whole trees were padded with black pixels with corresponding widths and heights so that the window slides over the bark area for an equivalent number of times. Then, the receptive window slid over the padded image with a predefined stride of half window size and calculated the CAM of the cropped image. All CAM calculations were then combined and averaged along the entire bark image.
Performance metrics
Two CNN models were evaluated using four performance metrics: overall accuracy and macro-averaged precision, recall, and f-score with \(\upbeta\) = 1. Each metric was calculated by the multi-class classification performance measurements adopted from Sokolova and Lapalme47, using overall accuracy instead of average accuracy in this study. The measurements were as follows:
where \(t{p}_{i}\), \({\text{tn}}_{i}\), \({fp}_{i}\), and \(f{n}_{i}\) represent true positive, true negative, false positive, and false negative counts, respectively, for each species class \({\mathrm{C}}_{\mathrm{i}}\) and with \(\mathrm{l}\) number of total classes (\(\mathrm{l}=42\)). \(\mathrm{M}\) indices represent the macro-averaged calculation of each metric.
Training details
When training VGG-16 and EfficientNet, we utilized the following hyper-parameters: 10−5 as the learning rate, Adaptive Moment Estimation (Adam) as the optimizer, 50% for dropout rate of the fully connected layer, and training batch size of 8. Two CNN models were pre-trained with the ImageNet dataset, and then fine-tuned until the maximum overall accuracy on test dataset did not increase for 10 consecutive epochs. CNN models were trained and tested under Windows 10 pro 64-bit OS and Python version 3.8.5, with hardware specification of an AMD Ryzen 5 3600XT CPU, 32 GB of RAM and an NVIDIA GTX 1080 Ti GPU. Tensorflow (2.4.1) and Keras (2.4.3) were utilized as deep learning frameworks.
Data availability
The BarkNet 1.0 dataset was retrieved from https://github.com/ulaval-damas/tree-bark-classification27; the bark image data collected in this study were published and are available on Zenodo (https://doi.org/10.5281/zenodo.4749062)48.
Code availability
The python scripts used in this study are available on GitHub (https://github.com/snutp/TBKFE). Refer to the scripts in the repository for technical implementation details of the methods in this study.
References
Mata-Montero, E. & Carranza-Rojas, J. Automated plant species identification: challenges and opportunities. in IFIP World Information Technology Forum, 26–36 (2016).
Hopkins, G. W. & Freckleton, R. P. Declines in the numbers of amateur and professional taxonomists: implications for conservation. Anim. Conserv. 5, 245–249. https://doi.org/10.1017/S1367943002002299 (2002).
Pearson, D. L., Hamilton, A. L. & Erwin, T. L. Recovery plan for the endangered taxonomy profession. Bioscience 61, 58–63. https://doi.org/10.1525/bio.2011.61.1.11 (2011).
Azlah, M. A. F., Chua, L. S., Rahmad, F. R., Abdullah, F. I. & Alwi, S. R. W. Review on techniques for plant leaf classification and recognition. Computers 8, 77. https://doi.org/10.3390/computers8040077 (2019).
Lima, M. C. F., Leandro, M. E. D. D., Valero, C., Coronel, L. C. P. & Bazzo, C. O. G. Automatic detection and monitoring of insect pests-a review. Agriculture-Basel 10, 161. https://doi.org/10.3390/agriculture10050161 (2020).
MacLeod, N., Benfield, M. & Culverhouse, P. Time to automate identification. Nature 467, 154–155. https://doi.org/10.1038/467154a (2010).
Norouzzadeh, M. S. et al. Automatically identifying, counting, and describing wild animals in camera-trap images with deep learning. Proc. Natl. Acad. Sci. USA 115, E5716–E5725. https://doi.org/10.1073/pnas.1719367115 (2018).
Rathi, D., Jain, S. & Indu, S. Underwater fish species classification using convolutional neural network and deep learning. 2017 Ninth international conference on advances in pattern recognition (ICAPR), 1–6 (2017).
Liang, Y.-Q., Li, J.-W., Li, J. & Valimaki, S. K. Impact of urbanization on plant diversity: A case study in built-up areas of Beijing. For. Stud. China 10, 179–188. https://doi.org/10.1007/s11632-008-0036-4 (2008).
Gogul, I. & Kumar, V. S. Flower species recognition system using convolution neural networks and transfer learning. in 2017 fourth international conference on signal processing, communication and networking (ICSCN), 1–6 (2017).
Kumar, N. et al. Leafsnap: A computer vision system for automatic plant species identification. in European conference on computer vision, 502–516 (2012).
Lee, S. H., Chan, C. S., Wilkin, P. & Remagnino, P. Deep-plant: Plant identification with convolutional neural networks. in 2015 IEEE international conference on image processing (ICIP), 452–456 (2015).
Nilsback, M.-E. & Zisserman, A. Automated flower classification over a large number of classes. in 2008 Sixth Indian Conference on Computer Vision, Graphics & Image Processing, 722–729 (2008).
Rzanny, M., Mader, P., Deggelmann, A., Chen, M. Q. & Waldchen, J. Flowers, leaves or both? How to obtain suitable images for automated plant identification. Plant Methods 15, 1–11. https://doi.org/10.1186/s13007-019-0462-4 (2019).
Wilf, P. et al. Computer vision cracks the leaf code. Proc. Natl. Acad. Sci. USA 113, 3305–3310. https://doi.org/10.1073/pnas.1524473113 (2016).
Cope, J. S., Corney, D., Clark, J. Y., Remagnino, P. & Wilkin, P. J. Plant species identification using digital morphometrics: A review. Expert Syst. Appl. 39, 7562–7573 (2012).
Ratajczak, R., Bertrand, S., Crispim-Junior, C., F & Tougne, L. Efficient Bark Recognition in the Wild. in International Conference on Computer Vision Theory and Applications (VISAPP 2019). https://doi.org/10.5220/0007361902400248 (2019).
Junikka, L. Survey of english macroscopic bark terminology. IAWA J. 15, 3–45. https://doi.org/10.1163/22941932-90001338 (1994).
Whitmore, T. C. Studies in systematic bark morphology I Bark Morphology in Dipterocarpaceae. New Phytol. 61, 191–207. https://doi.org/10.1111/j.1469-8137.1962.tb06288.x (1962).
Yunus, M., Yunus, D. & Iqbal, M. Systematic bark morphology of some tropical trees. Bot. J. Linn. Soc. 103, 367–377. https://doi.org/10.1111/j.1095-8339.1990.tb00196.x (1990).
Hadlich, H. L. et al. Recognizing Amazonian tree species in the field using bark tissues spectra. For. Ecol. Manage. 427, 296–304. https://doi.org/10.1016/j.foreco.2018.06.002 (2018).
Fekri-Ershad, S. Bark texture classification using improved local ternary patterns and multilayer neural network. Expert Syst. Appl. 158, 113509. https://doi.org/10.1016/j.eswa.2020.113509 (2020).
Chi, Z., Houqiang, L. & Chao, W. in International Conference on Neural Networks and Signal Processing, 2003. Proceedings of the 2003. 1035–1038 (IEEE).
Fiel, S. & Sablatnig, R. Automated identification of tree species from images of the bark , leaves or needles. in 16th Computer Vision Winter Workshop, 67–74 (2010).
Boudra, S., Yahiaoui, I. & Behloul, A. A comparison of multi-scale local binary pattern variants for bark image retrieval. in International conference on advanced concepts for intelligent vision systems, 764–775 (2015).
Krizhevsky, A., Sutskever, I. & Hinton, G. E. Imagenet classification with deep convolutional neural networks. Adv. Neural. Inf. Process. Syst. 25, 1097–1105 (2012).
Carpentier, M., Giguère, P. & Gaudreault, J. Tree species identification from bark images using convolutional neural networks. in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 1075–1081 (2018).
Dziugaite, G. K., Ben-David, S. & Roy, D. M. Enforcing Interpretability and its Statistical Impacts: Trade-offs between Accuracy and Interpretability. arXiv preprint arXiv:2010.13764 (2020).
Selvaraju, R. R. et al. Grad-CAM: Visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis. 128, 336–359. https://doi.org/10.1007/s11263-019-01228-7 (2020).
Pham, H., Dai, Z., Xie, Q. & Le, Q. V. Meta pseudo labels. in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 11557–11568 (2021).
Xie, Q., Luong, M.-T., Hovy, E. & Le, Q. V. Self-training with noisy student improves imagenet classification. in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 10687–10698 (2020).
Rosell, J. A., Gleason, S., Mendez-Alonzo, R., Chang, Y. & Westoby, M. Bark functional ecology: evidence for tradeoffs, functional coordination, and environment producing bark diversity. New Phytol. 201, 486–497. https://doi.org/10.1111/nph.12541 (2014).
Zhou, B., Khosla, A., Lapedriza, A., Oliva, A. & Torralba, A. Learning deep features for discriminative localization. in Proceedings of the IEEE conference on computer vision and pattern recognition, 2921–2929 (2016).
Chattopadhay, A., Sarkar, A., Howlader, P. & Balasubramanian, V. N. in 2018 IEEE Winter Conference on Applications of Computer Vision (WACV). 839–847 (IEEE).
Deng, J. et al. Imagenet: A large-scale hierarchical image database. in 2009 IEEE conference on computer vision and pattern recognition, 248–255 (2009).
LeCun, Y., Bengio, Y. & Hinton, G. Deep learning. Nature 521, 436–444. https://doi.org/10.1038/nature14539 (2015).
Zhang, Y. & Davison, B. D. Impact of imagenet model selection on domain adaptation. in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision Workshops, 173–182 (2020).
Kim, T. K. et al. Utilizing machine learning for detecting flowering in mid-range digital repeat photography. Ecol. Model. 440, 109419. https://doi.org/10.1016/j.ecolmodel.2020.109419 (2021).
Szegedy, C. et al. Going deeper with convolutions. in Proceedings of the IEEE conference on computer vision and pattern recognition, 1–9 (2015).
Mizoguchi, T., Ishii, A., Nakamura, H., Inoue, T. & Takamatsu, H. Lidar-based individual tree species classification using convolutional neural network. in Videometrics, Range Imaging, and Applications XIV 10332, 103320O (2017).
Cubuk, E. D., Zoph, B., Shlens, J. & Le, Q. V. Randaugment: Practical automated data augmentation with a reduced search space. in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 702–703 (2020).
Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014).
Tan, M. & Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. in International Conference on Machine Learning, 6105–6114 (2019).
Rangarajan, A. K. & Purushothaman, R. Disease classification in eggplant using pre-trained VGG16 and MSVM. Sci. Rep. 10, 1–11. https://doi.org/10.1038/s41598-020-59108-x (2020).
Baek, Y., Lee, B., Han, D., Yun, S. & Lee, H. Character region awareness for text detection. in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 9365–9374 (2019).
Majeed, Y., Karkee, M., Zhang, Q., Fu, L. S. & Whiting, M. D. Determining grapevine cordon shape for automated green shoot thinning using semantic segmentation-based deep learning networks. Comput. Electron. Agr. 171, 105308. https://doi.org/10.1016/j.compag.2020.105308 (2020).
Sokolova, M. & Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manage. 45, 427–437. https://doi.org/10.1016/j.ipm.2009.03.002 (2009).
Kim, T. K., Baek, G. H. & Kim, H. S. Tree bark identification dataset (BARK-KR). https://doi.org/10.5281/zenodo.4749062 (2021).
Acknowledgements
We are thankful for the support of the Seoul National University Arboretums (Gwanak, and Suwon) and their staffs.
Funding
This work was supported by Korea Forest Service (Korea Forestry Promotion Institute), and National Research Foundation of Korea (NRF), project No. 2018113B10-2020-BB01, No. 2020185D10-2222-AA02, and No. 2021R1A4A1025553.
Author information
Authors and Affiliations
Contributions
T.K.K.: conceptualization, data collection, visualization, methodology, software, validation, writing—original draft, writing—review and editing. J.H.: visualization, methodology, validation. D.R.: methodology, visualization. S.K.: methodology, visualization. S.B.: visualization. W.H.: writing—review and editing, visualization. K.K.: methodology, visualization. G.H.B.: data collection, methodology. H.S.K.: conceptualization, writing—review and editing, project administration, supervision, funding acquisition.
Corresponding author
Ethics declarations
Competing interests
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kim, T.K., Hong, J., Ryu, D. et al. Identifying and extracting bark key features of 42 tree species using convolutional neural networks and class activation mapping. Sci Rep 12, 4772 (2022). https://doi.org/10.1038/s41598-022-08571-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-022-08571-9
This article is cited by
-
A deep learning-based approach for the identification of selected species of genus Euphorbia L.
International Journal of Information Technology (2023)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.