索引于
  • 打开 J 门
  • Genamics 期刊搜索
  • 学术钥匙
  • 研究圣经
  • 宇宙IF
  • 访问全球在线农业研究 (AGORA)
  • 电子期刊图书馆
  • 参考搜索
  • 研究期刊索引目录 (DRJI)
  • 哈姆达大学
  • 亚利桑那州EBSCO
  • OCLC-WorldCat
  • 学者指导
  • SWB 在线目录
  • 虚拟生物学图书馆 (vifabio)
  • 普布隆斯
  • 日内瓦医学教育与研究基金会
  • 欧洲酒吧
  • 谷歌学术
分享此页面
期刊传单
Flyer image

抽象的

Explainable Convolutional Neural Network Based Tree Species Classification Using Multispectral Images from an Unmanned Aerial Vehicle

Ling-Wei Chen, Pin-Hui Lee, Yueh-Min Huang*

We seek to address labor shortages, in particular, the aging workforce of rural areas and thus facilitate agricultural management. The movement and operation of agricultural equipment in Taiwan is complicated by the fact that many commercial crops in Taiwan are planted on hillsides. For mixed crops in such sloped farming areas, the identification of tree species aids in agricultural management and reduces the labor needed for farming operations. General optical images collected by visible-light cameras are sufficient for recording but yield suboptimal results in tree species identification. Using a multispectral camera makes it possible to identify plants based on their spectral responses. We present a method for tree species classification using UAV visible light and multispectral imagery. We leverage the differences in spectral reflectance values between tree species and use near infrared band images to improve the model’s classification performance. CNN based deep neural models are widely used and yield highaccuracies, but 100% correct results are difficult to achieve, and model complexity generally increases with performance. This leads to uncertainty about the system’s final decisions. Interpretable AI extracts key information and interprets it to yield a better understanding of the model’s conclusions or actions.We use visualization(four pixel level attribution methods and one region level attribution method) to interpret the model post-hoc. Fuzzy IG for pixel level attribution best represents texture features, and region level attribution represents life regions more effectively than pixel level attribution, which aids human understanding.