TY - JOUR
T1 - StreetAware
T2 - A High-Resolution Synchronized Multimodal Urban Scene Dataset
AU - Piadyk, Yurii
AU - Rulff, Joao
AU - Brewer, Ethan
AU - Hosseini, Maryam
AU - Ozbay, Kaan
AU - Sankaradas, Murugan
AU - Chakradhar, Srimat
AU - Silva, Claudio
N1 - Funding Information:
This work was supported in part by CNS-1828576, CNS-1544753, CNS-1229185, CCF-1533564, CNS-1730396, and CNS-1626098. This study was partially supported by C2SMART, a Tier 1 USDOT University Transportation Center at New York University. Claudio Silva is partially supported by the DARPA PTG program. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of DARPA.
Publisher Copyright:
© 2023 by the authors.
PY - 2023/4
Y1 - 2023/4
N2 - Access to high-quality data is an important barrier in the digital analysis of urban settings, including applications within computer vision and urban design. Diverse forms of data collected from sensors in areas of high activity in the urban environment, particularly at street intersections, are valuable resources for researchers interpreting the dynamics between vehicles, pedestrians, and the built environment. In this paper, we present a high-resolution audio, video, and LiDAR dataset of three urban intersections in Brooklyn, New York, totaling almost 8 unique hours. The data were collected with custom Reconfigurable Environmental Intelligence Platform (REIP) sensors that were designed with the ability to accurately synchronize multiple video and audio inputs. The resulting data are novel in that they are inclusively multimodal, multi-angular, high-resolution, and synchronized. We demonstrate four ways the data could be utilized — (1) to discover and locate occluded objects using multiple sensors and modalities, (2) to associate audio events with their respective visual representations using both video and audio modes, (3) to track the amount of each type of object in a scene over time, and (4) to measure pedestrian speed using multiple synchronized camera views. In addition to these use cases, our data are available for other researchers to carry out analyses related to applying machine learning to understanding the urban environment (in which existing datasets may be inadequate), such as pedestrian-vehicle interaction modeling and pedestrian attribute recognition. Such analyses can help inform decisions made in the context of urban sensing and smart cities, including accessibility-aware urban design and Vision Zero initiatives.
AB - Access to high-quality data is an important barrier in the digital analysis of urban settings, including applications within computer vision and urban design. Diverse forms of data collected from sensors in areas of high activity in the urban environment, particularly at street intersections, are valuable resources for researchers interpreting the dynamics between vehicles, pedestrians, and the built environment. In this paper, we present a high-resolution audio, video, and LiDAR dataset of three urban intersections in Brooklyn, New York, totaling almost 8 unique hours. The data were collected with custom Reconfigurable Environmental Intelligence Platform (REIP) sensors that were designed with the ability to accurately synchronize multiple video and audio inputs. The resulting data are novel in that they are inclusively multimodal, multi-angular, high-resolution, and synchronized. We demonstrate four ways the data could be utilized — (1) to discover and locate occluded objects using multiple sensors and modalities, (2) to associate audio events with their respective visual representations using both video and audio modes, (3) to track the amount of each type of object in a scene over time, and (4) to measure pedestrian speed using multiple synchronized camera views. In addition to these use cases, our data are available for other researchers to carry out analyses related to applying machine learning to understanding the urban environment (in which existing datasets may be inadequate), such as pedestrian-vehicle interaction modeling and pedestrian attribute recognition. Such analyses can help inform decisions made in the context of urban sensing and smart cities, including accessibility-aware urban design and Vision Zero initiatives.
KW - computer vision
KW - data synchronization
KW - street-level imagery
KW - urban intelligence
KW - urban multimedia data
KW - urban sensing
UR - http://www.scopus.com/inward/record.url?scp=85152319874&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85152319874&partnerID=8YFLogxK
U2 - 10.3390/s23073710
DO - 10.3390/s23073710
M3 - Article
C2 - 37050773
AN - SCOPUS:85152319874
SN - 1424-8220
VL - 23
JO - Sensors
JF - Sensors
IS - 7
M1 - 3710
ER -