Abstract

Landscape architects and planners have been assessing eye-level vegetation to develop evidence-based designs, including the relationships between urban nature and human health. Measuring eye-level vegetation was often subjective and time-consuming in the past. Recent advances in computer vision have made it feasible to automatically measure eye-level greenery at a large scale. However, researchers still know little about the agreements of recent machine-based methods with human perception. The research gap may lead to inaccurate or even misleading findings that may prevent effective design and planning.This study tested the agreements between eye-level greenery detected by two machine-based methods (Brown Dog Green Index Extractor (BDGI) and PSP-Net) and human perception (manual selection via Photoshop Histogram). These two machine-based tools were selected because of their distinctive mechanisms: color detection and semantic segmentation. Cronbach’s alpha, correlation test, and Bland-Altman’s Plots were used to test agreements. Then, logistic regressions were used to find relationships between shades and vegetation density and the disagreement odds. Both tools closely agreed with human assessment in predicting eye-level greenery, with BDGI slightly closer to human. Vegetation density, but not percentage of shade, predicted the higher disagreement odds between PSP-Net and others. This finding will help advancing computer-based assessment of urban nature and contribute to our knowledge in assessing and linking eye-level greenery with potential outcomes such as physical and mental health and other design assessments.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call