Cargando…
Multimodal deep learning from satellite and street-level imagery for measuring income, overcrowding, and environmental deprivation in urban areas
Data collected at large scale and low cost (e.g. satellite and street level imagery) have the potential to substantially improve resolution, spatial coverage, and temporal frequency of measurement of urban inequalities. Multiple types of data from different sources are often available for a given ge...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
American Elsevier Pub. Co
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7985619/ https://www.ncbi.nlm.nih.gov/pubmed/33941991 http://dx.doi.org/10.1016/j.rse.2021.112339 |
_version_ | 1783668285329375232 |
---|---|
author | Suel, Esra Bhatt, Samir Brauer, Michael Flaxman, Seth Ezzati, Majid |
author_facet | Suel, Esra Bhatt, Samir Brauer, Michael Flaxman, Seth Ezzati, Majid |
author_sort | Suel, Esra |
collection | PubMed |
description | Data collected at large scale and low cost (e.g. satellite and street level imagery) have the potential to substantially improve resolution, spatial coverage, and temporal frequency of measurement of urban inequalities. Multiple types of data from different sources are often available for a given geographic area. Yet, most studies utilize a single type of input data when making measurements due to methodological difficulties in their joint use. We propose two deep learning-based methods for jointly utilizing satellite and street level imagery for measuring urban inequalities. We use London as a case study for three selected outputs, each measured in decile classes: income, overcrowding, and environmental deprivation. We compare the performances of our proposed multimodal models to corresponding unimodal ones using mean absolute error (MAE). First, satellite tiles are appended to street level imagery to enhance predictions at locations where street images are available leading to improvements in accuracy by 20, 10, and 9% in units of decile classes for income, overcrowding, and living environment. The second approach, novel to the best of our knowledge, uses a U-Net architecture to make predictions for all grid cells in a city at high spatial resolution (e.g. for 3 m × 3 m pixels in London in our experiments). It can utilize city wide availability of satellite images as well as more sparse information from street-level images where they are available leading to improvements in accuracy by 6, 10, and 11%. We also show examples of prediction maps from both approaches to visually highlight performance differences. |
format | Online Article Text |
id | pubmed-7985619 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | American Elsevier Pub. Co |
record_format | MEDLINE/PubMed |
spelling | pubmed-79856192021-05-01 Multimodal deep learning from satellite and street-level imagery for measuring income, overcrowding, and environmental deprivation in urban areas Suel, Esra Bhatt, Samir Brauer, Michael Flaxman, Seth Ezzati, Majid Remote Sens Environ Article Data collected at large scale and low cost (e.g. satellite and street level imagery) have the potential to substantially improve resolution, spatial coverage, and temporal frequency of measurement of urban inequalities. Multiple types of data from different sources are often available for a given geographic area. Yet, most studies utilize a single type of input data when making measurements due to methodological difficulties in their joint use. We propose two deep learning-based methods for jointly utilizing satellite and street level imagery for measuring urban inequalities. We use London as a case study for three selected outputs, each measured in decile classes: income, overcrowding, and environmental deprivation. We compare the performances of our proposed multimodal models to corresponding unimodal ones using mean absolute error (MAE). First, satellite tiles are appended to street level imagery to enhance predictions at locations where street images are available leading to improvements in accuracy by 20, 10, and 9% in units of decile classes for income, overcrowding, and living environment. The second approach, novel to the best of our knowledge, uses a U-Net architecture to make predictions for all grid cells in a city at high spatial resolution (e.g. for 3 m × 3 m pixels in London in our experiments). It can utilize city wide availability of satellite images as well as more sparse information from street-level images where they are available leading to improvements in accuracy by 6, 10, and 11%. We also show examples of prediction maps from both approaches to visually highlight performance differences. American Elsevier Pub. Co 2021-05 /pmc/articles/PMC7985619/ /pubmed/33941991 http://dx.doi.org/10.1016/j.rse.2021.112339 Text en © 2021 The Author(s) http://creativecommons.org/licenses/by/4.0/ This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Suel, Esra Bhatt, Samir Brauer, Michael Flaxman, Seth Ezzati, Majid Multimodal deep learning from satellite and street-level imagery for measuring income, overcrowding, and environmental deprivation in urban areas |
title | Multimodal deep learning from satellite and street-level imagery for measuring income, overcrowding, and environmental deprivation in urban areas |
title_full | Multimodal deep learning from satellite and street-level imagery for measuring income, overcrowding, and environmental deprivation in urban areas |
title_fullStr | Multimodal deep learning from satellite and street-level imagery for measuring income, overcrowding, and environmental deprivation in urban areas |
title_full_unstemmed | Multimodal deep learning from satellite and street-level imagery for measuring income, overcrowding, and environmental deprivation in urban areas |
title_short | Multimodal deep learning from satellite and street-level imagery for measuring income, overcrowding, and environmental deprivation in urban areas |
title_sort | multimodal deep learning from satellite and street-level imagery for measuring income, overcrowding, and environmental deprivation in urban areas |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7985619/ https://www.ncbi.nlm.nih.gov/pubmed/33941991 http://dx.doi.org/10.1016/j.rse.2021.112339 |
work_keys_str_mv | AT suelesra multimodaldeeplearningfromsatelliteandstreetlevelimageryformeasuringincomeovercrowdingandenvironmentaldeprivationinurbanareas AT bhattsamir multimodaldeeplearningfromsatelliteandstreetlevelimageryformeasuringincomeovercrowdingandenvironmentaldeprivationinurbanareas AT brauermichael multimodaldeeplearningfromsatelliteandstreetlevelimageryformeasuringincomeovercrowdingandenvironmentaldeprivationinurbanareas AT flaxmanseth multimodaldeeplearningfromsatelliteandstreetlevelimageryformeasuringincomeovercrowdingandenvironmentaldeprivationinurbanareas AT ezzatimajid multimodaldeeplearningfromsatelliteandstreetlevelimageryformeasuringincomeovercrowdingandenvironmentaldeprivationinurbanareas |