
Updates from our group
Full list of publications is here.

Greenery plays a vital role in urban environments, providing numerous benefits through diverse pathways. Various metrics and methodologies have been proposed to assess multiple dimensions of greenery exposure. For a comprehensive and precise assessment of greenery exposure for different research purposes, it is crucial to identify the most suitable methods and data sources. However, existing reviews primarily address the health outcomes of urban greenery, rather than the methods of assessing greenery exposure. To address this gap, we conducted a review of 312 research articles, focusing on methodologies and technologies for measuring greenery exposure in urban settings. This review categorizes exposure measurement techniques into three categories: proximity-based, mobility-based, and visibility-based, evaluating their strengths, limitations, and synergies. Proximity-based methods generally assess overall greenery level in residential areas or other locations, but they fall short in capturing the actual interactions between humans and greenery. Mobility-based methods track real-time human location and assess greenery exposure based on travel trajectories, but they neglect the specific nature of human-greenery interactions. In contrast, emerging visibility-based methods offer opportunities to measure potential visual interactions between individuals and greenery. We found emerging metrics tend to integrate 3D data, qualitative aspects, and diverse data sources. We advocate for an integrated approach that encompasses both human mobility and potential interactions with greenery across various areas. We also argue that data granularity is balanced against cost, scalability, and ethical constraints. Our comprehensive review offers a framework and categorization to guide studies in designing exposure measurements aligned with their research objectives.

Three-dimensional urban environment simulation is a powerful tool for informed urban planning. However, the intensive manual effort required to prepare input 3D city models has hindered its widespread adoption. To address this challenge, we present VoxCity, an open-source Python package that provides a one-stop solution for grid-based 3D city model generation and urban environment simulation for cities worldwide. VoxCity’s ‘generator’ subpackage automatically downloads building heights, tree canopy heights, land cover, and terrain elevation within a specified target area, and voxelizes buildings, trees, land cover, and terrain to generate an integrated voxel city model. The ‘simulator’ subpackage enables users to conduct environmental simulations, including solar radiation and view index analyses. Users can export the generated models using several file formats compatible with external software, such as ENVI-met (INX), Blender, and Rhino (OBJ). We generated 3D city models for eight global cities, and demonstrated the calculation of solar irradiance, sky view index, and green view index. We also showcased microclimate simulation and 3D rendering visualization through ENVI-met and Rhino, respectively, through the file export function. Additionally, we reviewed openly available geospatial data to create guidelines to help users choose appropriate data sources depending on their target areas and purposes. VoxCity can significantly reduce the effort and time required for 3D city model preparation and promote the utilization of urban environment simulations. This contributes to more informed urban and architectural design that considers environmental impacts, and in turn, fosters sustainable and livable cities. VoxCity is released openly at https://github.com/kunifujiwara/VoxCity.

Building properties, such as height, usage, and material, play a crucial role in spatial data infrastructures, supporting various urban applications. Despite their importance, comprehensive building attribute data remain scarce in many urban areas. Recent advances have enabled the extraction of objective building attributes using remote sensing and street-level imagery. However, establishing a pipeline that integrates diverse open datasets, acquires holistic building imagery, and infers comprehensive building attributes at scale remains a significant challenge. Among the first, this study bridges the gaps by introducing OpenFACADES, an open framework that leverages multimodal crowdsourced data to enrich building profiles with both objective attributes and semantic descriptors through multimodal large language models. First, we integrate street-level image metadata from Mapillary with OpenStreetMap geometries via isovist analysis, identifying images that provide suitable vantage points for observing target buildings. Second, we automate the detection of building facades in panoramic imagery and tailor a reprojection approach to convert objects into holistic perspective views that approximate real-world observation. Third, we introduce an innovative approach that harnesses and investigates the capabilities of open-source large vision-language models (VLMs) for multi-attribute prediction and open-vocabulary captioning in building-level analytics, leveraging a globally sourced dataset of 31,180 labeled images from seven cities. Evaluation shows that fine-tuned VLM excel in multi-attribute inference, outperforming single-attribute computer vision models and zero-shot ChatGPT-4o. Further experiments confirm its superior generalization and robustness across culturally distinct region and varying image conditions. Finally, the model is applied for large-scale building annotation, generating a dataset of 1.2 million images for half a million buildings. This open‐source framework enhances the scope, adaptability, and granularity of building‐level assessments, enabling more fine‐grained and interpretable insights into the built environment. Our dataset and code are available openly at: https://github.com/seshing/OpenFACADES.

Understanding people’s preferences is crucial for urban planning, yet current approaches often combine responses from multi-cultural populations, obscuring demographic differences and risking amplifying biases. We conducted a large-scale urban visual perception survey of streetscapes worldwide using street view imagery, examining how demographics—including gender, age, income, education, race and ethnicity, and personality traits—shape perceptions among 1,000 participants with balanced demographics from five countries and 45 nationalities. This dataset, Street Perception Evaluation Considering Socioeconomics, reveals demographic- and personality-based differences across six traditional indicators—safe, lively, wealthy, beautiful, boring, depressing—and four new ones: live nearby, walk, cycle, green. Location-based sentiments further shape these preferences. Machine-learning models trained on existing global datasets tend to overestimate positive indicators and underestimate negative ones compared to human responses, underscoring the need for local context. Our study aspires to rectify the myopic treatment of street perception, which rarely considers demographics or personality traits.

The 15-minute city concept emphasizes accessible urban living by ensuring essential services are reachable within walking or biking distance. However, most evaluations rely on two-dimensional (2D) analyses, neglecting the vertical complexity of high-density cities. This study introduces a 3D framework for assessing 15-minute accessibility in Nanjing, China. Using natural language processing and rule-based methods, we construct a 3D functional composition dataset from multi-source data. We then develop floor-level proximity indices that account for both horizontal travel time and vertical circulation (e.g., stairs, elevators). Analyzing over 90 million simulated trips, we find that accessibility generally declines with building height, though access to offices and commercial facilities improves at 20th or higher floors. Spatial inequalities emerge not only between central and peripheral zones but also across building levels and regional GDP levels, with a U-shaped disparity tied to distance from downtown. Notably, 11%–17% of trips considered accessible in 2D analyses exceed the 15-minute threshold when vertical travel is included. Our findings highlight the need to incorporate vertical space in 15-minute city evaluations and offer a scalable method to support inclusive, fair, and livable 3D urban planning with the background of 15-minute city.