Comparison of automatic and interactive thresholding of hemispherical photography

A. Inoue, K. Yamamoto, N. Mizoue

Research output: Contribution to journalArticlepeer-review

10 Citations (Scopus)


This study presents the effects of operator bias and variation in interactive thresholding on the estimation of light environment using hemispherical photography. Twenty-one hemispherical photographs taken beneath a wide range of canopy densities were visually converted to binary images twice by 21 operators, and then the gap fraction was computed from the images. The interactive threshold varied greatly among the different operators and within a single operator, which resulted in a considerable operator bias and variation in the gap fraction. This study also compared three widely used automatic thresholding algorithms, which were installed in freely available software LIA for Win32 for analyzing hemispherical photography, with interactive thresholding using the same photographs. The median of the interactive threshold by repetitive interactive thresholdings from 21 operators was assumed to be correct for the comparison. The results indicated that MINIMUM was considered to be a better algorithm than the other ones installed in LIA32 when the gap fraction was over 10%. However, VARIANCE seemed to be superior to MINIMUM under the low gap fraction and the cloudy sky condition with dark and white clouds. This implied that MINIMUM or VARIANCE should be used for analyzing hemispherical photographs with LIA32. In conclusion, we need to pay attention to the selection of the automatic thresholding algorithm and the sky condition when taking hemispherical photographs.

Original languageEnglish
Pages (from-to)78-87
Number of pages10
JournalJournal of Forest Science
Issue number2
Publication statusPublished - Feb 2011

All Science Journal Classification (ASJC) codes

  • Forestry
  • Soil Science


Dive into the research topics of 'Comparison of automatic and interactive thresholding of hemispherical photography'. Together they form a unique fingerprint.

Cite this