Dataset Open Access

# A Convolutional Neural Network classifier identifies tree species in mixed-conifer forest from hyperspectral imagery

Geoffrey A Fricker; Jonathan Daniel Ventura; Jeffrey Wolf; Malcolm P. North; Frank W. Davis; Janet Franklin

### DataCite XML Export

<?xml version='1.0' encoding='utf-8'?>
<identifier identifierType="DOI">10.5281/zenodo.3470250</identifier>
<creators>
<creator>
<creatorName>Geoffrey A Fricker</creatorName>
<affiliation>Social Sciences Department, California Polytechnic State University, San Luis Obispo</affiliation>
</creator>
<creator>
<creatorName>Jonathan Daniel Ventura</creatorName>
<affiliation>Department of Computer Sciences and Software Engineering, California Polytechnic State University, San Luis Obispo</affiliation>
</creator>
<creator>
<creatorName>Jeffrey Wolf</creatorName>
<affiliation>Amazon Corporation</affiliation>
</creator>
<creator>
<creatorName>Malcolm P. North</creatorName>
<affiliation>US Forest Service, PSW Research Station</affiliation>
</creator>
<creator>
<creatorName>Frank W. Davis</creatorName>
<affiliation>Bren School of Environmental Science and Management, University of California, Santa Barbara</affiliation>
</creator>
<creator>
<creatorName>Janet Franklin</creatorName>
<affiliation>Department of Botany and Plant Sciences, University of California, Riverside</affiliation>
</creator>
</creators>
<titles>
<title>A Convolutional Neural Network classifier identifies tree species in mixed-conifer forest from hyperspectral imagery</title>
</titles>
<publisher>Zenodo</publisher>
<publicationYear>2019</publicationYear>
<subjects>
<subject>shapefile</subject>
<subject>tree species</subject>
<subject>convolutional neural network</subject>
<subject>hyperspectral imagery</subject>
<subject>NEON</subject>
</subjects>
<dates>
<date dateType="Issued">2019-09-27</date>
</dates>
<language>en</language>
<resourceType resourceTypeGeneral="Dataset"/>
<alternateIdentifiers>
<alternateIdentifier alternateIdentifierType="url">https://zenodo.org/record/3470250</alternateIdentifier>
</alternateIdentifiers>
<relatedIdentifiers>
<relatedIdentifier relatedIdentifierType="DOI" relationType="IsVersionOf">10.5281/zenodo.3463588</relatedIdentifier>
</relatedIdentifiers>
<rightsList>
<rights rightsURI="info:eu-repo/semantics/openAccess">Open Access</rights>
</rightsList>
<descriptions>
<description descriptionType="Abstract">&lt;p&gt;Published online:&amp;nbsp;&lt;a href="https://www.mdpi.com/2072-4292/11/19/2326"&gt;https://www.mdpi.com/2072-4292/11/19/2326&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;DOI: 10.3390/rs11192326&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Abstract:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In this study, we automate tree species classification and mapping using field-based training data, high spatial resolution airborne hyperspectral imagery, and a convolutional neural network classifier (CNN). We tested our methods by identifying seven dominant trees species as well as dead standing trees in a mixed-conifer forest in the Southern Sierra Nevada Mountains, CA (USA) using training, validation, and testing datasets composed of spatially-explicit transects and plots sampled across a single strip of imaging spectroscopy. We also used a three-band &amp;lsquo;Red-Green-Blue&amp;rsquo; pseudo true-color subset of the hyperspectral imagery strip to test the classification accuracy of a CNN model without the additional non-visible spectral data provided in the hyperspectral imagery. Our classifier is pixel-based rather than object based, although we use three-dimensional structural information from airborne Light Detection and Ranging (LiDAR) to identify trees (points &amp;gt; 5 m above the ground) and the classifier was applied to image pixels that were thus identified as tree crowns. By training a CNN classifier using field data and hyperspectral imagery, we were able to accurately identify tree species and predict their distribution, as well as the distribution of tree mortality, across the landscape. Using a window size of 15 pixels and eight hidden convolutional layers, a CNN model classified the correct species of 713 individual trees from hyperspectral imagery with an average F-score of 0.87 and F-scores ranging from 0.67&amp;ndash;0.95 depending on species. The CNN classification model performance increased from a combined F-score of 0.64 for the Red-Green-Blue model to a combined F-score of 0.87 for the hyperspectral model. The hyperspectral CNN model captures the species composition changes across ~700 meters (1935 to 2630 m) of elevation from a lower-elevation mixed oak conifer forest to a higher-elevation fir-dominated coniferous forest. High resolution tree species maps can support forest ecosystem monitoring and management, and identifying dead trees aids landscape assessment of forest mortality resulting from drought, insects and pathogens. We publicly provide our code to apply deep learning classifiers to tree species identification from geospatial imagery and field training data&lt;/p&gt;

&lt;p&gt;Digital Publication of the training data polygons and hyperspectral imagery used in&amp;nbsp;the manuscript &amp;quot;A Convolutional Neural Network classifier identifies tree species in mixed-conifer forest from hyperspectral imagery&amp;quot;.&lt;/p&gt;

&lt;p&gt;Code is available in a Jupyter Notebook and can be found here:&amp;nbsp;&lt;a href="https://github.com/jonathanventura/canopy"&gt;https://github.com/jonathanventura/canopy&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;National Ecological Observatory Network. 2018. Provisional data downloaded from&amp;nbsp;&lt;a href="http://data.neonscience.org/"&gt;http://data.neonscience.org&lt;/a&gt;&amp;nbsp;on 22&amp;nbsp;June 2018. Battelle, Boulder, CO, USA&lt;/p&gt;</description>
<description descriptionType="Other">Data to replicate the experiment is available for download in two zipped files:
"NEON_D17_TEAK_DP1QA_20170627_181333_RGB_Reflectance.zip" (Imagery)
"CNN_LABELS_2019.zip" (Training Label Shapefiles)

* Note: The imagery is 5.5 gb (zipped).

All code used to run the analysis is located in a repository here:
https://github.com/jonathanventura/canopy

The only flightline you will need to repeat our results is called "NEON_D17_TEAK_DP1_20170627_181333".
If you download your own NEON data, the raw HDF 5 files can be converted to a geotiff using R code found here:
http://neonscience.github.io/neon-data-institute-2016//R/open-NEON-hdf5-functions/

Contact the National Ecological Observatory Network (NEON) to download the comparable imagery data files for all sites and collections: https://data.neonscience.org/home.</description>
</descriptions>
</resource>

408
6,510
views